Deep Nonparametric Estimation of Operators between Infinite Dimensional Spaces

Hao Liu, Haizhao Yang*, Minshuo Chen, Tuo Zhao, Wenjing Liao*

*Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

Abstract

Learning operators between infinitely dimensional spaces is an important learning task arising in machine learning, imaging science, mathematical modeling and simulations, etc. This paper studies the nonparametric estimation of Lipschitz operators using deep neural networks. Non-asymptotic upper bounds are derived for the generalization error of the empirical risk minimizer over a properly chosen network class. Under the assumption that the target operator exhibits a low dimensional structure, our error bounds decay as the training sample size increases, with an attractive fast rate depending on the intrinsic dimension in our estimation. Our assumptions cover most scenarios in real applications and our results give rise to fast rates by exploiting low dimensional structures of data in operator estimation. We also investigate the influence of network structures (e.g., network width, depth, and sparsity) on the generalization error of the neural network estimator and propose a general suggestion on the choice of network structures to maximize the learning efficiency quantitatively.
Original languageEnglish
Article number24
Pages (from-to)1-67
Number of pages67
JournalJournal of Machine Learning Research
Volume25
DOIs
Publication statusPublished - 24 Jan 2024

User-Defined Keywords

  • Deep neural networks
  • Generalization error analysis
  • Nonparametric estimation
  • Operator learning

Fingerprint

Dive into the research topics of 'Deep Nonparametric Estimation of Operators between Infinite Dimensional Spaces'. Together they form a unique fingerprint.

Cite this