TY - JOUR
T1 - Distributed Gradient Descent for Functional Learning
AU - Yu, Zhan
AU - Fan, Jun
AU - Shi, Zhongjie
AU - Zhou, Ding Xuan
N1 - The work described in this paper was partially supported by InnoHK initiative, The Government of the HKSAR, Laboratory for AI-Powered Financial Technologies, the Research Grants Council of Hong Kong [Project No. # CityU 11308121 and No. N_CityU102/20], and National Natural Science Foundation of China under [Project No. 11461161006]. The work by Jun Fan is partially supported by the Research Grants Council of Hong Kong [Project No. HKBU 12302819] and [Project No. HKBU 12301619].
Publisher Copyright:
© 2024 IEEE
PY - 2024/9
Y1 - 2024/9
N2 - In recent years, different types of distributed and parallel learning schemes have received increasing attention for their strong advantages in handling large-scale data information. In the information era, to face the big data challenges that stem from functional data analysis very recently, we propose a novel distributed gradient descent functional learning (DGDFL) algorithm to tackle functional data across numerous local machines (processors) in the framework of reproducing kernel Hilbert space. Based on integral operator approaches, we provide the first theoretical understanding of the DGDFL algorithm in many different aspects of the literature. On the way of understanding DGDFL, firstly, a data-based gradient descent functional learning (GDFL) algorithm associated with a single-machine model is proposed and comprehensively studied. Under mild conditions, confidence-based optimal learning rates of DGDFL are obtained without the saturation boundary on the regularity index suffered in previous works in functional regression. We further provide a semi-supervised DGDFL approach to weaken the restriction on the maximal number of local machines to ensure optimal rates. To our best knowledge, the DGDFL provides the first divide-and-conquer iterative training approach to functional learning based on data samples of intrinsically infinite-dimensional random functions (functional covariates) and enriches the methodologies for functional data analysis.
AB - In recent years, different types of distributed and parallel learning schemes have received increasing attention for their strong advantages in handling large-scale data information. In the information era, to face the big data challenges that stem from functional data analysis very recently, we propose a novel distributed gradient descent functional learning (DGDFL) algorithm to tackle functional data across numerous local machines (processors) in the framework of reproducing kernel Hilbert space. Based on integral operator approaches, we provide the first theoretical understanding of the DGDFL algorithm in many different aspects of the literature. On the way of understanding DGDFL, firstly, a data-based gradient descent functional learning (GDFL) algorithm associated with a single-machine model is proposed and comprehensively studied. Under mild conditions, confidence-based optimal learning rates of DGDFL are obtained without the saturation boundary on the regularity index suffered in previous works in functional regression. We further provide a semi-supervised DGDFL approach to weaken the restriction on the maximal number of local machines to ensure optimal rates. To our best knowledge, the DGDFL provides the first divide-and-conquer iterative training approach to functional learning based on data samples of intrinsically infinite-dimensional random functions (functional covariates) and enriches the methodologies for functional data analysis.
KW - divide and conquer
KW - functional data
KW - functional linear model
KW - gradient descent
KW - intergal operator
KW - learning theory
KW - reproducing kernel
KW - semi-supervised learning
UR - https://ieeexplore.ieee.org/document/10599555
UR - http://www.scopus.com/inward/record.url?scp=85198726225&partnerID=8YFLogxK
U2 - 10.1109/TIT.2024.3428325
DO - 10.1109/TIT.2024.3428325
M3 - Journal article
AN - SCOPUS:85198726225
SN - 0018-9448
VL - 70
SP - 6547
EP - 6571
JO - IEEE Transactions on Information Theory
JF - IEEE Transactions on Information Theory
IS - 9
ER -