Title:
Distributed learning and inference in deep models

dc.contributor.advisor Fekri, Faramarz
dc.contributor.author Abdi, Afshin
dc.contributor.committeeMember AlRegib, Ghassan
dc.contributor.committeeMember Romberg, Justin
dc.contributor.committeeMember Bloch, Matthieu
dc.contributor.committeeMember Maguluri, Siva Theja
dc.contributor.department Electrical and Computer Engineering
dc.date.accessioned 2020-09-08T12:48:03Z
dc.date.available 2020-09-08T12:48:03Z
dc.date.created 2020-08
dc.date.issued 2020-07-22
dc.date.submitted August 2020
dc.date.updated 2020-09-08T12:48:04Z
dc.description.abstract In recent years, the size of deep learning problems has been increased significantly, both in terms of the number of available training samples as well as the number of parameters and complexity of the model. In this thesis, we considered the challenges encountered in training and inference of large deep models, especially on nodes with limited computational power and capacity. We studied two classes of related problems; 1) distributed training of deep models, and 2) compression and restructuring of deep models for efficient distributed and parallel execution to reduce inference times. Especially, we considered the communication bottleneck in distributed training and inference of deep models. Data compression is a viable tool to mitigate the communication bottleneck in distributed deep learning. However, the existing methods suffer from a few drawbacks, such as the increased variance of stochastic gradients (SG), slower convergence rate, or added bias to SG. In my Ph.D. research, we have addressed these challenges from three different perspectives: 1) Information Theory and the CEO Problem, 2) Indirect SG compression via Matrix Factorization, and 3) Quantized Compressive Sampling. We showed, both theoretically and via simulations, that our proposed methods can achieve smaller MSE than other unbiased compression methods with fewer communication bit-rates, resulting in superior convergence rates. Next, we considered federated learning over wireless multiple access channels (MAC). Efficient communication requires the communication algorithm to satisfy the constraints imposed by the nodes in the network and the communication medium. To satisfy these constraints and take advantage of the over-the-air computation inherent in MAC, we proposed a framework based on random linear coding and developed efficient power management and channel usage techniques to manage the trade-offs between power consumption and communication bit-rate. In the second part of this thesis, we considered the distributed parallel implementation of an already-trained deep model on multiple workers. Since latency due to the synchronization and data transfer among workers adversely affects the performance of the parallel implementation, it is desirable to have minimum interdependency among parallel sub-models on the workers. To achieve this goal, we developed and analyzed RePurpose, an efficient algorithm to rearrange the neurons in the neural network and partition them (without changing the general topology of the neural network) such that the interdependency among sub-models is minimized under the computations and communications constraints of the workers.
dc.description.degree Ph.D.
dc.format.mimetype application/pdf
dc.identifier.uri http://hdl.handle.net/1853/63671
dc.language.iso en_US
dc.publisher Georgia Institute of Technology
dc.subject Machine learning
dc.subject Artificial intelligence
dc.subject Distributed training
dc.subject Distributed learning
dc.title Distributed learning and inference in deep models
dc.type Text
dc.type.genre Dissertation
dspace.entity.type Publication
local.contributor.advisor Fekri, Faramarz
local.contributor.corporatename School of Electrical and Computer Engineering
local.contributor.corporatename College of Engineering
relation.isAdvisorOfPublication f46b53e9-5ee4-4646-b71c-8bd223cd31c8
relation.isOrgUnitOfPublication 5b7adef2-447c-4270-b9fc-846bd76f80f2
relation.isOrgUnitOfPublication 7c022d60-21d5-497c-b552-95e489a06569
thesis.degree.level Doctoral
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
ABDI-DISSERTATION-2020.pdf
Size:
3.6 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
LICENSE.txt
Size:
3.86 KB
Format:
Plain Text
Description: