TY - JOUR
T1 - Enhancing Collaborative Inference on Heterogeneous Edge Devices via Adaptive Ensemble Knowledge Distillation
AU - Wu, Shangrui
AU - Li, Yupeng
AU - Wang, Wenhua
AU - Guo, Jianxiong
AU - Fan, Wentao
AU - Liu, Qin
AU - Jia, Weijia
AU - Yu, Shui
AU - Cao, Jiannong
AU - Wang, Tian
N1 - Publisher Copyright:
© 1983-2012 IEEE.
Funding Information:
This work was supported in part by grants from the National Natural Science Foundation of China (NSFC) (No. 62172046, No. 62372047, No. 62202402, No. 62276106), the Beijing Natural Science Foundation (No. 4232028), the Natural Science Foundation of Guangdong Province (No.
2024A1515011323), Zhuhai Basic and Applied Basic Research Foundation (No. 2220004002619), Higher Education Research Topics of Guangdong Association of Higher Education in the 14th Five-Year Plan under 24GYB207, Beijing Normal University Education Reform Project under jx2024139, the Shaanxi Key Laboratory of Information Communication Network and Security under Grant ICNS202102, the Youth Innovation Talent Project for Ordinary Universities of Department of Education of Guangdong Province under Grant 2023KQNCX151, Guangdong Basic and Applied Basic Research Foundation (No. 2023A1515011562), the Guangdong Provincial Key Laboratory of IRADS (No. 2022B1212010006), Hong Kong RGC Early Career Scheme (No. 22202423), the Initiation Grant for Faculty Niche Research Areas 2023/24 (No. RC-FNRA-IG/23-24/COMM/01).
PY - 2025/5/28
Y1 - 2025/5/28
N2 - The integration of edge computing with deep neural networks (DNNs) is
crucial for intelligent industrial cyber-physical systems. Typically,
deploying DNNs on heterogeneous edge devices relies on methods like
model compression and partitioning. However, these approaches often
result in homogeneous models across devices. This homogeneity limits the
collective capability of edge computing systems, particularly in terms
of generalization to diverse data distributions and adaptation to
dynamic industrial environments. In this work, we propose to treat each
DNN on an edge device as an independent model, aggregating their
capabilities via ensemble learning to enhance generalization and dynamic
adaptability. To realize this, we introduce the Adaptive Ensemble
Knowledge Distillation Framework (AEKDF), combining cloud-based model
training with edge computing based collaborative inference. In the
cloud, AEKDF develops an enhanced Born Again Network that generates
diverse, lightweight models tailored to specific edge devices through
knowledge distillation. This process ensures model diversity which is
critical to effective ensemble learning. On the edge, AEKDF employs an
adaptive ensemble technique that aggregates prediction logits across
devices, enabling rapid adaptation to changing environments and
maintaining inference efficiency. Our extensive evaluations conducted on
a realistic prototype demonstrate the substantial boost in predictive
performance achieved by our AEKDF, showcasing a 4% to 10% accuracy
improvement on the CIFAR-100 compared to conventional single-model
approaches, while maintaining low latency.
AB - The integration of edge computing with deep neural networks (DNNs) is
crucial for intelligent industrial cyber-physical systems. Typically,
deploying DNNs on heterogeneous edge devices relies on methods like
model compression and partitioning. However, these approaches often
result in homogeneous models across devices. This homogeneity limits the
collective capability of edge computing systems, particularly in terms
of generalization to diverse data distributions and adaptation to
dynamic industrial environments. In this work, we propose to treat each
DNN on an edge device as an independent model, aggregating their
capabilities via ensemble learning to enhance generalization and dynamic
adaptability. To realize this, we introduce the Adaptive Ensemble
Knowledge Distillation Framework (AEKDF), combining cloud-based model
training with edge computing based collaborative inference. In the
cloud, AEKDF develops an enhanced Born Again Network that generates
diverse, lightweight models tailored to specific edge devices through
knowledge distillation. This process ensures model diversity which is
critical to effective ensemble learning. On the edge, AEKDF employs an
adaptive ensemble technique that aggregates prediction logits across
devices, enabling rapid adaptation to changing environments and
maintaining inference efficiency. Our extensive evaluations conducted on
a realistic prototype demonstrate the substantial boost in predictive
performance achieved by our AEKDF, showcasing a 4% to 10% accuracy
improvement on the CIFAR-100 compared to conventional single-model
approaches, while maintaining low latency.
KW - Deep Neural Networks
KW - Edge Computing
KW - Ensemble Learning
KW - Industrial Internet of Things
UR - https://www.scopus.com/pages/publications/105006919365
U2 - 10.1109/JSAC.2025.3574594
DO - 10.1109/JSAC.2025.3574594
M3 - Journal article
AN - SCOPUS:105006919365
SN - 0733-8716
VL - 43
SP - 1
EP - 16
JO - IEEE Journal on Selected Areas in Communications
JF - IEEE Journal on Selected Areas in Communications
IS - 10
ER -