Title:
Theoretical Error Performance Analysis for Deep Neural Network Based Regression Functional Approximation

dc.contributor.advisor Lee, Chin-Hui
dc.contributor.advisor Ma, Xiaoli
dc.contributor.author Qi, Jun
dc.contributor.committeeMember Anderson, David
dc.contributor.committeeMember Romberg, Justin
dc.contributor.committeeMember Siniscalchi, Sabato
dc.contributor.department Electrical and Computer Engineering
dc.date.accessioned 2022-01-14T16:19:52Z
dc.date.available 2022-01-14T16:19:52Z
dc.date.created 2021-12
dc.date.issued 2022-01-12
dc.date.submitted December 2021
dc.date.updated 2022-01-14T16:19:52Z
dc.description.abstract Based on Kolmogorov's superposition theorem and universal approximation theorems by Cybenko and Barron, any vector-to-scalar function can be approximated by a multi-layer perceptron (MLP) within certain bounds. The theorems inspire us to exploit deep neural networks (DNN) based vector-to-vector regression. This dissertation aims at establishing theoretical foundations on DNN based vector-to-vector functional approximation, and bridging the gap between DNN based applications and their theoretical understanding in terms of representation and generalization powers. Concerning the representation power, we develop the classical universal approximation theorems and put forth a new upper bound to vector-to-vector regression. More specifically, we first derive upper bounds on the artificial neural network (ANN), and then we generalize the concepts to DNN based architectures. Our theorems suggest that a broader width of the top hidden layer and a deep model structure bring a more expressive power of DNN based vector-to-vector regression, which is illustrated with speech enhancement experiments. As for the generalization power of DNN based vector-to-vector regression, we employ a well-known error decomposition technique, which factorizes an expected loss into the sum of an approximation error, an estimation error, and an optimization error. Since the approximation error is associated with our attained upper bound upon the expressive power, we concentrate our research on deriving the upper bound for the estimation error and optimization error based on statistical learning theory and non-convex optimization. Moreover, we demonstrate that mean absolute error (MAE) satisfies the property of Lipschitz continuity and exhibits better performance than mean squared error (MSE). The speech enhancement experiments with DNN models are utilized to corroborate our aforementioned theorems. Finally, since an over-parameterized setting for DNN is expected to ensure our theoretical upper bounds on the generalization power, we put forth a novel deep tensor learning framework, namely tensor-train deep neural network (TT-DNN), to deal with an explosive DNN model size and realize effective deep regression with much smaller model complexity. Our experiments of speech enhancement demonstrate that a TT-DNN can maintain or even achieve higher performance accuracy but with much fewer model parameters than an even over-parameterized DNN.
dc.description.degree Ph.D.
dc.format.mimetype application/pdf
dc.identifier.uri http://hdl.handle.net/1853/66188
dc.language.iso en_US
dc.publisher Georgia Institute of Technology
dc.subject deep neural network
dc.subject vector-to-vector regression
dc.subject error decomposition
dc.subject representation power
dc.subject generalization power
dc.title Theoretical Error Performance Analysis for Deep Neural Network Based Regression Functional Approximation
dc.type Text
dc.type.genre Dissertation
dspace.entity.type Publication
local.contributor.advisor Ma, Xiaoli
local.contributor.advisor Lee, Chin-Hui
local.contributor.corporatename School of Electrical and Computer Engineering
local.contributor.corporatename College of Engineering
relation.isAdvisorOfPublication 82df2e0c-c5d7-455c-88ed-e9fb64e26407
relation.isAdvisorOfPublication b35c0c49-bee2-49ad-9d6a-867b4ba8908b
relation.isOrgUnitOfPublication 5b7adef2-447c-4270-b9fc-846bd76f80f2
relation.isOrgUnitOfPublication 7c022d60-21d5-497c-b552-95e489a06569
thesis.degree.level Doctoral
Files
Original bundle
Now showing 1 - 1 of 1
Thumbnail Image
Name:
QI-DISSERTATION-2021.pdf
Size:
5.16 MB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
LICENSE.txt
Size:
3.86 KB
Format:
Plain Text
Description: