Abstract
The Influence Function (IF) is a widely used technique for assessing the impact of individual training samples on model predictions. However, existing IF methods often fail to provide reliable influence estimates in deep neural networks, particularly when applied to noisy training data. This issue does not stem from inaccuracies in parameter change estimation, which has been the primary focus of prior research, but rather from deficiencies in loss change estimation, specifically due to the sharpness of validation risk. In this work, we establish a theoretical connection between influence estimation error, validation set risk, and its sharpness, underscoring the importance of flat validation minima for accurate influence estimation. Furthermore, we introduce a novel estimation form of Influence Function specifically designed for flat validation minima. Experimental results across various tasks validate the superiority of our approach.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of the 42nd International Conference on Machine Learning, ICML 2025 |
| Publisher | ML Research Press |
| Pages | 72091-72111 |
| Number of pages | 21 |
| Publication status | Published - Jul 2025 |
| Event | 42nd International Conference on Machine Learning, ICML 2025 - Vancouver Convention Center, Vancouver, Canada Duration: 13 Jul 2025 → 19 Jul 2025 https://icml.cc/Conferences/2025 (Conference Website) https://icml.cc/virtual/2025/calendar (Conference Calendar) https://proceedings.mlr.press/v267/ (Conference Proceedings) |
Publication series
| Name | Proceedings of Machine Learning Research |
|---|---|
| Publisher | ML Research Press |
| Volume | 267 |
Conference
| Conference | 42nd International Conference on Machine Learning, ICML 2025 |
|---|---|
| Country/Territory | Canada |
| City | Vancouver |
| Period | 13/07/25 → 19/07/25 |
| Internet address |
|