Understanding and Improving Feature Learning for Out-of-Distribution Generalization

Yongqiang Chen, Wei Huang, Kaiwen Zhou, Yatao Bian, Bo Han, James Cheng

Research output: Chapter in book/report/conference proceedingConference proceedingpeer-review

7 Citations (Scopus)

Abstract

A common explanation for the failure of out-of-distribution (OOD) generalization is that the model trained with empirical risk minimization (ERM) learns spurious features instead of invariant features.However, several recent studies challenged this explanation and found that deep networks may have already learned sufficiently good features for OOD generalization.Despite the contradictions at first glance, we theoretically show that ERM essentially learns both spurious and invariant features, while ERM tends to learn spurious features faster if the spurious correlation is stronger.Moreover, when fed the ERM learned features to the OOD objectives, the invariant feature learning quality significantly affects the final OOD performance, as OOD objectives rarely learn new features.Therefore, ERM feature learning can be a bottleneck to OOD generalization.To alleviate the reliance, we propose Feature Augmented Training (FeAT), to enforce the model to learn richer features ready for OOD generalization.FeAT iteratively augments the model to learn new features while retaining the already learned features.In each round, the retention and augmentation operations are performed on different subsets of the training data that capture distinct features.Extensive experiments show that FeAT effectively learns richer features thus boosting the performance of various OOD objectives.

Original languageEnglish
Title of host publicationAdvances in Neural Information Processing Systems 36 (NeurIPS 2023)
EditorsA. Oh, T. Naumann, A. Globerson, K. Saenko, M. Hardt, S. Levine
PublisherNeural Information Processing Systems Foundation
Number of pages55
ISBN (Print)9781713899921
DOIs
Publication statusPublished - 10 Dec 2023
Event37th Conference on Neural Information Processing Systems, NeurIPS 2023 - Ernest N. Morial Convention Center, New Orleans, United States
Duration: 10 Dec 202316 Dec 2023
https://proceedings.neurips.cc/paper_files/paper/2023 (Conference Paper Search)
https://openreview.net/group?id=NeurIPS.cc/2023/Conference#tab-accept-oral (Conference Paper Search)
https://neurips.cc/Conferences/2023 (Conference Website)

Publication series

NameAdvances in Neural Information Processing Systems
PublisherNeural information processing systems foundation
Volume36
ISSN (Print)1049-5258

Conference

Conference37th Conference on Neural Information Processing Systems, NeurIPS 2023
Country/TerritoryUnited States
CityNew Orleans
Period10/12/2316/12/23
Internet address

Scopus Subject Areas

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'Understanding and Improving Feature Learning for Out-of-Distribution Generalization'. Together they form a unique fingerprint.

Cite this