TY - GEN
T1 - Proximal average approximated incremental gradient method for composite penalty regularized empirical risk minimization
AU - Cheung, Yiu Ming
AU - Lou, Jian
N1 - This work was supported by the Faculty Research Grant of Hong Kong Baptist University under Projects: FRG2/14-15/075 and FRG1/14-15/041, and the National Science Foundation of China under Grant: 61272366.
Publisher Copyright:
© 2015 Y.-m. Cheung & J. Lou.
PY - 2015/11
Y1 - 2015/11
N2 - Proximal average (PA) is an approximation technique proposed recently to handle non-smooth composite regularizer in empirical risk minimization problem. For nonsmooth composite regularizer, it is often difficult to directly derive the corresponding proximal update when solving with popular proximal update. While traditional approaches resort to complex splitting methods like ADMM, proximal average provides an alternative, featuring the tractability of implementation and theoretical analysis. Nevertheless, compared to SDCA-ADMM and SAG-ADMM which are examples of ADMM-based methods achieving faster convergence rate and low per-iteration complexity, existing PA-based approaches either converge slowly (e.g. PA-ASGD) or suffer from high per-iteration cost (e.g. PA-APG). In this paper, we therefore propose a new PA-based algorithm called PA-SAGA, which is optimal in both convergence rate and per-iteration cost, by incorporating into incremental gradient-based framework.
AB - Proximal average (PA) is an approximation technique proposed recently to handle non-smooth composite regularizer in empirical risk minimization problem. For nonsmooth composite regularizer, it is often difficult to directly derive the corresponding proximal update when solving with popular proximal update. While traditional approaches resort to complex splitting methods like ADMM, proximal average provides an alternative, featuring the tractability of implementation and theoretical analysis. Nevertheless, compared to SDCA-ADMM and SAG-ADMM which are examples of ADMM-based methods achieving faster convergence rate and low per-iteration complexity, existing PA-based approaches either converge slowly (e.g. PA-ASGD) or suffer from high per-iteration cost (e.g. PA-APG). In this paper, we therefore propose a new PA-based algorithm called PA-SAGA, which is optimal in both convergence rate and per-iteration cost, by incorporating into incremental gradient-based framework.
UR - http://proceedings.mlr.press/v45/Cheung15.html
UR - http://www.scopus.com/inward/record.url?scp=85026309768&partnerID=8YFLogxK
M3 - Conference proceeding
AN - SCOPUS:85026309768
T3 - Proceedings of Machine Learning Research
SP - 205
EP - 220
BT - 7th Asian Conference on Machine Learning, ACML 2015
A2 - Holmes, Geoffrey
A2 - Liu, Tie-Yan
PB - ML Research Press
T2 - 7th Asian Conference on Machine Learning, ACML 2015
Y2 - 20 November 2015 through 22 November 2015
ER -