TY - JOUR
T1 - Deep Nonparametric Estimation of Operators between Infinite Dimensional Spaces
AU - Liu, Hao
AU - Yang, Haizhao
AU - Chen, Minshuo
AU - Zhao, Tuo
AU - Liao, Wenjing
N1 - Publisher copyright:
© 2024 Hao Liu, Haizhao Yang, Minshuo Chen, Tuo Zhao and Wenjing Liao.
PY - 2024/1/24
Y1 - 2024/1/24
N2 - Learning operators between infinitely dimensional spaces is an important learning task arising in machine learning, imaging science, mathematical modeling and simulations, etc. This paper studies the nonparametric estimation of Lipschitz operators using deep neural networks. Non-asymptotic upper bounds are derived for the generalization error of the empirical risk minimizer over a properly chosen network class. Under the assumption that the target operator exhibits a low dimensional structure, our error bounds decay as the training sample size increases, with an attractive fast rate depending on the intrinsic dimension in our estimation. Our assumptions cover most scenarios in real applications and our results give rise to fast rates by exploiting low dimensional structures of data in operator estimation. We also investigate the influence of network structures (e.g., network width, depth, and sparsity) on the generalization error of the neural network estimator and propose a general suggestion on the choice of network structures to maximize the learning efficiency quantitatively.
AB - Learning operators between infinitely dimensional spaces is an important learning task arising in machine learning, imaging science, mathematical modeling and simulations, etc. This paper studies the nonparametric estimation of Lipschitz operators using deep neural networks. Non-asymptotic upper bounds are derived for the generalization error of the empirical risk minimizer over a properly chosen network class. Under the assumption that the target operator exhibits a low dimensional structure, our error bounds decay as the training sample size increases, with an attractive fast rate depending on the intrinsic dimension in our estimation. Our assumptions cover most scenarios in real applications and our results give rise to fast rates by exploiting low dimensional structures of data in operator estimation. We also investigate the influence of network structures (e.g., network width, depth, and sparsity) on the generalization error of the neural network estimator and propose a general suggestion on the choice of network structures to maximize the learning efficiency quantitatively.
KW - Deep neural networks
KW - Generalization error analysis
KW - Nonparametric estimation
KW - Operator learning
UR - https://www.jmlr.org/papers/v25/22-0719.html
UR - https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=hkbuirimsintegration2023&SrcAuth=WosAPI&KeyUT=WOS:001168520600001&DestLinkType=FullRecord&DestApp=WOS_CPL
M3 - Journal article
SN - 1532-4435
VL - 25
SP - 1
EP - 67
JO - Journal of Machine Learning Research
JF - Journal of Machine Learning Research
M1 - 24
ER -