Title:
Explicit Group Sparse Projection for Machine Learning

dc.contributor.advisor Calhoun, Vince D.
dc.contributor.advisor Plis, Sergey
dc.contributor.author Ohib, Riyasat
dc.contributor.committeeMember Bloch, Matthieu
dc.contributor.committeeMember Anderson, David
dc.contributor.department Electrical and Computer Engineering
dc.date.accessioned 2024-01-10T18:53:40Z
dc.date.available 2024-01-10T18:53:40Z
dc.date.created 2023-12
dc.date.issued 2023-12-14
dc.date.submitted December 2023
dc.date.updated 2024-01-10T18:53:41Z
dc.description.abstract The concept of sparse solutions in classical machine learning is noted for its efficiency and has parallels in the natural world, such as in the mammalian visual cortex. This biological basis provides an inspiration for the importance of sparsity in computational models. Sparsity is increasingly relevant in machine learning, especially in non-negative matrix factorization (NMF), where it aids in interpretability and efficiency. NMF involves breaking down a non-negative matrix into simpler components, with sparsity ensuring these components distinctly represent data features, simplifying interpretation. In deep learning, sparse model parameters lead to more efficient computation, quicker training and inference, and in some cases, more robust models. As models grow in size, the role of inducing sparsity becomes even more crucial. In this thesis, we design a new sparse projection method for a set of vectors that guarantees a desired average sparsity level measured leveraging the popular Hoyer measure. Existing approaches either project each vector individually or require the use of a regularization parameter which implicitly maps to the average $\ell_0$-measure of sparsity. Instead, in our approach we set the \revise{Hoyer} sparsity level for the whole set explicitly and simultaneously project a group of vectors with the \revise{Hoyer} sparsity level of each vector tuned automatically. Hence, we call this the Group Sparse Projection (GSP). We show that the computational complexity of our projection operator is linear in the size of the problem. GSP can be used in particular to sparsify the columns of a matrix, which we use to compute sparse low-rank matrix approximations (namely, sparse NMF). We showcase the efficacy of our approach in both supervised and unsupervised learning tasks on image datasets including MNIST and CIFAR10. In non-negative matrix factorization, our approach yields competitive reconstruction errors against state-of-the-art algorithms. In neural network pruning, the sparse models produced by our method have competitive accuracy at corresponding sparsity values compared to existing methods.
dc.description.degree M.S.
dc.format.mimetype application/pdf
dc.identifier.uri https://hdl.handle.net/1853/73220
dc.language.iso en_US
dc.publisher Georgia Institute of Technology
dc.subject Sparse Projection
dc.subject Sparsity
dc.subject sparse NMF
dc.subject optimization
dc.subject Neural Network Pruning
dc.subject Model Compression
dc.title Explicit Group Sparse Projection for Machine Learning
dc.type Text
dc.type.genre Thesis
dspace.entity.type Publication
local.contributor.advisor Calhoun, Vince D.
local.contributor.corporatename College of Engineering
local.contributor.corporatename School of Electrical and Computer Engineering
relation.isAdvisorOfPublication a8679c01-9c03-494e-80bf-489cfa53277a
relation.isOrgUnitOfPublication 7c022d60-21d5-497c-b552-95e489a06569
relation.isOrgUnitOfPublication 5b7adef2-447c-4270-b9fc-846bd76f80f2
thesis.degree.level Masters
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
OHIB-THESIS-2023.pdf
Size:
590.31 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
LICENSE.txt
Size:
3.86 KB
Format:
Plain Text
Description: