TY - JOUR
T1 - Stochastic AUC optimization with general loss
AU - Yang, Zhenhuan
AU - Shen, Wei
AU - Ying, Yiming
AU - Yuan, Xiaoming
N1 - Funding information:
This work was completed when Wei Shen was a visiting student at SUNY Albany. Yiming Ying is supported by the National Science Foundation (NSF, Grant IIS1816227)
Publisher Copyright:
© 2020 American Institute of Mathematical Sciences. All rights reserved.
PY - 2020/8
Y1 - 2020/8
N2 - Recently, there is considerable work on developing efficient stochastic optimization algorithms for AUC maximization. However, most of them focus on the least square loss which may be not the best option in practice. The main difficulty for dealing with the general convex loss is the pairwise nonlinearity w.r.t. the sampling distribution generating the data. In this paper, we use Bernstein polynomials to uniformly approximate the general losses which are able to decouple the pairwise nonlinearity. In particular, we show that this reduction for AUC maximization with a general loss is equivalent to a weakly convex (nonconvex) min-max formulation. Then, we develop a novel SGD algorithm for AUC maximization with per-iteration cost linearly w.r.t. the data dimension, making it amenable for streaming data analysis. Despite its non-convexity, we prove its global convergence by exploring the appealing convexity-preserving property of Bernstein polynomials and the intrinsic structure of the min-max formulation. Experiments are performed to validate the effectiveness of the proposed approach.
AB - Recently, there is considerable work on developing efficient stochastic optimization algorithms for AUC maximization. However, most of them focus on the least square loss which may be not the best option in practice. The main difficulty for dealing with the general convex loss is the pairwise nonlinearity w.r.t. the sampling distribution generating the data. In this paper, we use Bernstein polynomials to uniformly approximate the general losses which are able to decouple the pairwise nonlinearity. In particular, we show that this reduction for AUC maximization with a general loss is equivalent to a weakly convex (nonconvex) min-max formulation. Then, we develop a novel SGD algorithm for AUC maximization with per-iteration cost linearly w.r.t. the data dimension, making it amenable for streaming data analysis. Despite its non-convexity, we prove its global convergence by exploring the appealing convexity-preserving property of Bernstein polynomials and the intrinsic structure of the min-max formulation. Experiments are performed to validate the effectiveness of the proposed approach.
KW - AUC maximization
KW - Bernstein polynomial
KW - Stochastic optimization
UR - http://www.scopus.com/inward/record.url?scp=85090783542&partnerID=8YFLogxK
U2 - 10.3934/cpaa.2020188
DO - 10.3934/cpaa.2020188
M3 - Journal article
AN - SCOPUS:85090783542
SN - 1534-0392
VL - 19
SP - 4191
EP - 4212
JO - Communications on Pure and Applied Analysis
JF - Communications on Pure and Applied Analysis
IS - 8
ER -