Abstract
Given data with noisy labels, over-parameterized deep networks can gradually memorize the data, and fit everything in the end. Although equipped with corrections for noisy labels, many learning methods in this area still suffer overfitting due to undesired memorization. In this paper, to relieve this issue, we propose stochastic integrated gradient underweighted ascent (SIGUA): in a minibatch, we adopt gradient descent on good data as usual, and learning-rate-reduced gradient ascent on bad data; the proposal is a versatile approach where data goodness or badness is w.r.t. desired or undesired memorization given a base learning method. Technically, SIGUA pulls optimization back for generalization when their goals conflict with each other; philosophically, SIGUA shows forgetting undesired memorization can reinforce desired memorization. Experiments demonstrate that SIGUA successfully robustifies two typical base learning methods, so that their performance is often significantly improved.
Original language | English |
---|---|
Title of host publication | Proceedings of the 37th International Conference on Machine Learning, ICML 2020 |
Editors | Hal Daumé III, Aarti Singh |
Publisher | ML Research Press |
Pages | 3964-3974 |
Number of pages | 11 |
ISBN (Electronic) | 9781713821120 |
Publication status | Published - Jul 2020 |
Event | 37th International Conference on Machine Learning, ICML 2020 - Virtual, Online Duration: 13 Jul 2020 → 18 Jul 2020 https://proceedings.mlr.press/v119/ |
Publication series
Name | Proceedings of Machine Learning Research |
---|---|
Volume | 119 |
ISSN (Print) | 2640-3498 |
Conference
Conference | 37th International Conference on Machine Learning, ICML 2020 |
---|---|
Period | 13/07/20 → 18/07/20 |
Internet address |
Scopus Subject Areas
- Computational Theory and Mathematics
- Human-Computer Interaction
- Software