TY - JOUR
T1 - Learning to Incentivize
T2 - Convergence-Guaranteed Federated Learning Via Client Quality Discovery
AU - Wang, Jiajun
AU - Guo, Jianxiong
AU - Wang, Juncheng
AU - Ding, Xingjian
AU - Li, Deying
AU - Wu, Weili
N1 - This work was supported in part by the National Natural Science Foundation of China (NSFC) under Grant No. 62202016 and No. 62202055, the Guangdong Basic and Applied Basic Research Foundation under Grant No. 2025A1515012843, the Start-up Fund from Beijing Normal University under Grant No. 312200502510, the Internal Fund from Beijing Normal-Hong Kong Baptist University under Grant No. UICR0400003-24 and UICR0200022- 25, and the Interdisciplinary Intelligence Super Computer Center of Beijing Normal University (Zhuhai).
Publisher Copyright:
© 2025 IEEE
PY - 2025/12/4
Y1 - 2025/12/4
N2 - Federated learning (FL) is a privacy-preserving distributed machine learning framework where multiple devices collaborate with the assistance of an aggregator. However, the limitations of aggregator communication result in only a portion of clients with high data quality being selected to participate in FL, but the quality of clients’ data cannot be evaluated without access to the original data. Most existing methods for selecting clients employ a data quality metric with empirically defined scores, which may select clients with high non-ID degrees, thereby reducing the accuracy and freshness of the model. Furthermore, due to the unknown quality of clients’ data, current incentive mechanisms lack FL convergence guarantees, which prevent the client behavior from improving the global model accuracy. To address these issues, in this paper, we propose using the gradient difference as a metric for the quality of clients’ data, which can quantify the non-IID degree and contribution potential of each client. We formulate a client selection problem using the Combinatorial Multi-Armed Bandit (CMAB) model and design an effective selection strategy, improving the worst-case regret proof to provide a theoretical guarantee for it. Based on these results, we develop an incentive mechanism by the FL convergence analysis, quantifying the utility functions of the aggregator and clients, and modeling their interaction as a two-stage Stackelberg game. For the non-convex utility function, our method establishes the existence and uniqueness of the Stackelberg equilibrium, thereby enabling the determination of the optimal strategy for maximizing the utility of all participants. Finally, extensive simulation experiments on real-world datasets demonstrate the effectiveness of our proposed method compared to state-of-the-art approaches.
AB - Federated learning (FL) is a privacy-preserving distributed machine learning framework where multiple devices collaborate with the assistance of an aggregator. However, the limitations of aggregator communication result in only a portion of clients with high data quality being selected to participate in FL, but the quality of clients’ data cannot be evaluated without access to the original data. Most existing methods for selecting clients employ a data quality metric with empirically defined scores, which may select clients with high non-ID degrees, thereby reducing the accuracy and freshness of the model. Furthermore, due to the unknown quality of clients’ data, current incentive mechanisms lack FL convergence guarantees, which prevent the client behavior from improving the global model accuracy. To address these issues, in this paper, we propose using the gradient difference as a metric for the quality of clients’ data, which can quantify the non-IID degree and contribution potential of each client. We formulate a client selection problem using the Combinatorial Multi-Armed Bandit (CMAB) model and design an effective selection strategy, improving the worst-case regret proof to provide a theoretical guarantee for it. Based on these results, we develop an incentive mechanism by the FL convergence analysis, quantifying the utility functions of the aggregator and clients, and modeling their interaction as a two-stage Stackelberg game. For the non-convex utility function, our method establishes the existence and uniqueness of the Stackelberg equilibrium, thereby enabling the determination of the optimal strategy for maximizing the utility of all participants. Finally, extensive simulation experiments on real-world datasets demonstrate the effectiveness of our proposed method compared to state-of-the-art approaches.
KW - Client selection
KW - convergence
KW - data quality metric
KW - federated learning
KW - incentive mechanism
KW - multi-armed bandit
KW - stackelberg game
UR - https://www.scopus.com/pages/publications/105024128249
UR - https://ieeexplore.ieee.org/document/11278058/authors#authors
U2 - 10.1109/TMC.2025.3640183
DO - 10.1109/TMC.2025.3640183
M3 - Journal article
AN - SCOPUS:105024128249
SN - 1536-1233
JO - IEEE Transactions on Mobile Computing
JF - IEEE Transactions on Mobile Computing
ER -