TY - JOUR
T1 - A Time-consistency Curriculum for Learning from Instance-dependent Noisy Labels
AU - Wu, Songhua
AU - Zhou, Tianyi
AU - Du, Yuxuan
AU - Yu, Jun
AU - Han, Bo
AU - Liu, Tongliang
N1 - The work of Songhua Wu and Tongliang Liu were supported in part by Australian Research Council Project under Grant DE-190101473 and in part by RIKEN Collaborative Research Fund. The work of Tongliang Liu was supported in part by the following Australian Research Council projects under Grant FT220100318, Grant DP220102121, Grant LP220100527, Grant LP220200949, and Grant IC190100031. The work of Jun Yu was supported in part by the Natural Science Foundation of China under Grant 62276242, in part by National Aviation Science Foundation under Grant 2022Z071078001, in part by CAAI-Huawei MindSpore Open Fund under Grant CAAIXSJLJJ-2021-016B and Grant CAAIXSJLJJ-2022-001A, in part by Anhui Province Key Research and Development Program under Grant 202104a05020007, in part by Dreams Foundation of Jianghuai Advance Technology Center under Grant 2023-ZM01Z001, in part by Beijing Municipal Science & Technology Commission, Administrative Commission of Zhongguancun Science Park under Grant Z231100005923035. The work of Bo Han was supported by the RGC Early Career Scheme under Grant 22200720, in part by the NSFC Young Scientists Fund under Grant 62006202, in part by NSFC General Program under Grant 62376235, in part by Guangdong Basic and Applied Basic Research Foundation under Grant 2022A1515011652, in part by HKBU Faculty Niche Research Areas under Grant RC-FNRA-IG/22-23/SCI/04, and in part by HKBU CSD Departmental Incentive Scheme. This work was supported by the National Natural Science Foundation of China under Grant 61725105, Grant 62076241, Grant 62171436, and Grant 62101371, and in part by the Jiangsu Province Science Foundation for Youths under Grant BK20210707.
PY - 2024/7
Y1 - 2024/7
N2 - Many machine learning algorithms are known to be fragile on simple instance-independent noisy labels. However, noisy labels in real-world data are more devastating since they are produced by more complicated mechanisms in an instance-dependent manner. In this paper, we target this practical challenge of Instance-Dependent Noisy Labels by jointly training (1) a model reversely engineering the noise generating mechanism, which produces an instance-dependent mapping between the clean label posterior and the observed noisy label and (2) a robust classifier that produces clean label posteriors. Compared to previous methods, the former model is novel and enables end-to-end learning of the latter directly from noisy labels. An extensive empirical study indicates that the time-consistency of data is critical to the success of training both models and motivates us to develop a curriculum selecting training data based on their dynamics on the two models’ outputs over the course of training. We show that the curriculum-selected data provide both clean labels and high-quality input-output pairs for training the two models. Therefore, it leads to promising and robust classification performance even in notably challenging settings of instance-dependent noisy labels where many SoTA methods could easily fail. Extensive experimental comparisons and ablation studies further demonstrate the advantages and significance of the time-consistency curriculum in learning from instance-dependent noisy labels on multiple benchmark datasets.
AB - Many machine learning algorithms are known to be fragile on simple instance-independent noisy labels. However, noisy labels in real-world data are more devastating since they are produced by more complicated mechanisms in an instance-dependent manner. In this paper, we target this practical challenge of Instance-Dependent Noisy Labels by jointly training (1) a model reversely engineering the noise generating mechanism, which produces an instance-dependent mapping between the clean label posterior and the observed noisy label and (2) a robust classifier that produces clean label posteriors. Compared to previous methods, the former model is novel and enables end-to-end learning of the latter directly from noisy labels. An extensive empirical study indicates that the time-consistency of data is critical to the success of training both models and motivates us to develop a curriculum selecting training data based on their dynamics on the two models’ outputs over the course of training. We show that the curriculum-selected data provide both clean labels and high-quality input-output pairs for training the two models. Therefore, it leads to promising and robust classification performance even in notably challenging settings of instance-dependent noisy labels where many SoTA methods could easily fail. Extensive experimental comparisons and ablation studies further demonstrate the advantages and significance of the time-consistency curriculum in learning from instance-dependent noisy labels on multiple benchmark datasets.
KW - Instance-dependent noisy labels
KW - image classification
KW - time-consistent curriculum learning
UR - http://www.scopus.com/inward/record.url?scp=85184319351&partnerID=8YFLogxK
U2 - 10.1109/TPAMI.2024.3360623
DO - 10.1109/TPAMI.2024.3360623
M3 - Journal article
SN - 0162-8828
VL - 46
SP - 4830
EP - 4842
JO - IEEE Transactions on Pattern Analysis and Machine Intelligence
JF - IEEE Transactions on Pattern Analysis and Machine Intelligence
IS - 7
ER -