TY - JOUR
T1 - Winning Prize Comes from Losing Tickets
T2 - Improve Invariant Learning by Exploring Variant Parameters for Out-of-Distribution Generalization
AU - Huang, Zhuo
AU - Li, Muyang
AU - Shen, Li
AU - Yu, Jun
AU - Gong, Chen
AU - Han, Bo
AU - Liu, Tongliang
N1 - Li Shen is supported by STI 2030-Major Projects (No. 2021ZD0201405). Jun Yu is supported by supported by the Natural Science Foundation of China (62276242), National Aviation Science Foundation (2022Z071078001), CAAI-Huawei MindSpore Open Fund (CAAIXSJLJJ-2021-016B, CAAIXSJLJJ-2022-001A), Anhui Province Key Research and Development Program (202104a05020007), Dreams Foundation of Jianghuai Advance Technology Center (2023-ZM01Z001), USTC-IAT Application Sci. & Tech. Achievement Cultivation Program (JL06521001Y). Chen Gong is supported by NSF of China (No: 61973162), NSF of Jiangsu Province (No: BZ2021013), NSF for Distinguished Young Scholar of Jiangsu Province (No: BK20220080), and CAAI-Huawei MindSpore Open Fund. Bo Han is supported by the NSFC General Program No. 62376235, and Guangdong Basic and Applied Basic Research Foundation Nos. 2024A1515012399 and 2022A1515011652. Tongliang Liu is partially supported by the following Australian Research Council projects: FT220100318, DP220102121, LP220100527, LP220200949, and IC190100031.
Publisher Copyright:
© The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
PY - 2024/8/1
Y1 - 2024/8/1
N2 - Out-of-Distribution (OOD) Generalization aims to learn robust models that generalize well to various environments without fitting to distribution-specific features. Recent studies based on Lottery Ticket Hypothesis (LTH) address this problem by minimizing the learning target to find some of the parameters that are critical to the task. However, in open-world visual recognition problems, such solutions are suboptimal as the learning task contains severe distribution noises, which can mislead the optimization process. Therefore, apart from finding the task-related parameters (i.e., invariant parameters), we propose Exploring Variant parameters for Invariant Learning (EVIL) which also leverages the distribution knowledge to find the parameters that are sensitive to distribution shift (i.e., variant parameters). Once the variant parameters are left out of invariant learning, a robust subnetwork that is resistant to distribution shift can be found. Additionally, the parameters that are relatively stable across distributions can be considered invariant ones to improve invariant learning. By fully exploring both variant and invariant parameters, our EVIL can effectively identify a robust subnetwork to improve OOD generalization. In extensive experiments on integrated testbed: DomainBed, EVIL can effectively and efficiently enhance many popular methods, such as ERM, IRM, SAM, etc. Our code is available at https://github.com/tmllab/EVIL.
AB - Out-of-Distribution (OOD) Generalization aims to learn robust models that generalize well to various environments without fitting to distribution-specific features. Recent studies based on Lottery Ticket Hypothesis (LTH) address this problem by minimizing the learning target to find some of the parameters that are critical to the task. However, in open-world visual recognition problems, such solutions are suboptimal as the learning task contains severe distribution noises, which can mislead the optimization process. Therefore, apart from finding the task-related parameters (i.e., invariant parameters), we propose Exploring Variant parameters for Invariant Learning (EVIL) which also leverages the distribution knowledge to find the parameters that are sensitive to distribution shift (i.e., variant parameters). Once the variant parameters are left out of invariant learning, a robust subnetwork that is resistant to distribution shift can be found. Additionally, the parameters that are relatively stable across distributions can be considered invariant ones to improve invariant learning. By fully exploring both variant and invariant parameters, our EVIL can effectively identify a robust subnetwork to improve OOD generalization. In extensive experiments on integrated testbed: DomainBed, EVIL can effectively and efficiently enhance many popular methods, such as ERM, IRM, SAM, etc. Our code is available at https://github.com/tmllab/EVIL.
KW - Invariant learning
KW - Out-of-distribution generalization
KW - Sparse training
UR - http://www.scopus.com/inward/record.url?scp=85200134085&partnerID=8YFLogxK
U2 - 10.1007/s11263-024-02075-x
DO - 10.1007/s11263-024-02075-x
M3 - Journal article
AN - SCOPUS:85200134085
SN - 0920-5691
VL - 133
SP - 456
EP - 474
JO - International Journal of Computer Vision
JF - International Journal of Computer Vision
IS - 1
ER -