Generalization Guarantee of SGD for Pairwise Learning

Yunwen Lei, Mingrui Liu, Yiming Ying

Research output: Chapter in book/report/conference proceedingConference proceedingpeer-review

24 Citations (Scopus)

Abstract

Recently, there is a growing interest in studying pairwise learning since it includes many important machine learning tasks as specific examples, e.g., metric learning, AUC maximization and ranking. While stochastic gradient descent (SGD) is an efficient method, there is a lacking study on its generalization behavior for pairwise learning. In this paper, we present a systematic study on the generalization analysis of SGD for pairwise learning to understand the balance between generalization and optimization. We develop a novel high-probability generalization bound for uniformly-stable algorithms to incorporate the variance information for better generalization, based on which we establish the first nonsmooth learning algorithm to achieve almost optimal high-probability and dimension-independent excess risk bounds with O(n) gradient computations. We consider both convex and nonconvex pairwise learning problems. Our stability analysis for convex problems shows how the interpolation can help generalization. We establish a uniform convergence of gradients, and apply it to derive the first excess risk bounds on population gradients for nonconvex pairwise learning. Finally, we extend our stability analysis to pairwise learning with gradient-dominated problems.

Original languageEnglish
Title of host publication35th Conference on Neural Information Processing Systems (NeurIPS 2021)
EditorsMarc'Aurelio Ranzato, Alina Beygelzimer, Yann Dauphin, Percy S. Liang, Jenn Wortman Vaughan
PublisherNeural Information Processing Systems Foundation
Pages21216-21228
Number of pages13
Volume26
ISBN (Print)9781713845393
Publication statusPublished - 6 Dec 2021
Event35th Conference on Neural Information Processing Systems, NeurIPS 2021 - Virtual
Duration: 6 Dec 202114 Dec 2021
https://nips.cc/Conferences/2021 (Conference website)
https://neurips.cc/Conferences/2021 (Conference website)
https://papers.nips.cc/paper_files/paper/2021 (Conference proceedings)
https://proceedings.neurips.cc/paper/2021 (Conference proceedings)

Publication series

NameAdvances in Neural Information Processing Systems
Volume34
ISSN (Print)1049-5258
NameNeurIPS Proceedings

Conference

Conference35th Conference on Neural Information Processing Systems, NeurIPS 2021
Period6/12/2114/12/21
Internet address

Scopus Subject Areas

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'Generalization Guarantee of SGD for Pairwise Learning'. Together they form a unique fingerprint.

Cite this