TY - JOUR
T1 - Efficient methods for estimating constrained parameters with applications to regularized (lasso) logistic regression
AU - Tian, Guo Liang
AU - Tang, Man Lai
AU - Fang, Hong Bin
AU - Tan, Ming
N1 - Funding Information:
G.-L. Tian and M. Tan’s research was supported in part by U.S. National Cancer Institute grants CA119758 and CA106767. The work of M.-L. Tang was fully supported by Hong Kong Baptist University grant FRG/06-07/II-20. The research of H.-B. Fang was partially supported by U.S. National Cancer Institute grant CA106767.
PY - 2008/3/15
Y1 - 2008/3/15
N2 - Fitting logistic regression models is challenging when their parameters are restricted. In this article, we first develop a quadratic lower-bound (QLB) algorithm for optimization with box or linear inequality constraints and derive the fastest QLB algorithm corresponding to the smallest global majorization matrix. The proposed QLB algorithm is particularly suited to problems to which the EM-type algorithms are not applicable (e.g., logistic, multinomial logistic, and Cox's proportional hazards models) while it retains the same EM ascent property and thus assures the monotonic convergence. Secondly, we generalize the QLB algorithm to penalized problems in which the penalty functions may not be totally differentiable. The proposed method thus provides an alternative algorithm for estimation in lasso logistic regression, where the convergence of the existing lasso algorithm is not generally ensured. Finally, by relaxing the ascent requirement, convergence speed can be further accelerated. We introduce a pseudo-Newton method that retains the simplicity of the QLB algorithm and the fast convergence of the Newton method. Theoretical justification and numerical examples show that the pseudo-Newton method is up to 71 (in terms of CPU time) or 107 (in terms of number of iterations) times faster than the fastest QLB algorithm and thus makes bootstrap variance estimation feasible. Simulations and comparisons are performed and three real examples (Down syndrome data, kyphosis data, and colon microarray data) are analyzed to illustrate the proposed methods.
AB - Fitting logistic regression models is challenging when their parameters are restricted. In this article, we first develop a quadratic lower-bound (QLB) algorithm for optimization with box or linear inequality constraints and derive the fastest QLB algorithm corresponding to the smallest global majorization matrix. The proposed QLB algorithm is particularly suited to problems to which the EM-type algorithms are not applicable (e.g., logistic, multinomial logistic, and Cox's proportional hazards models) while it retains the same EM ascent property and thus assures the monotonic convergence. Secondly, we generalize the QLB algorithm to penalized problems in which the penalty functions may not be totally differentiable. The proposed method thus provides an alternative algorithm for estimation in lasso logistic regression, where the convergence of the existing lasso algorithm is not generally ensured. Finally, by relaxing the ascent requirement, convergence speed can be further accelerated. We introduce a pseudo-Newton method that retains the simplicity of the QLB algorithm and the fast convergence of the Newton method. Theoretical justification and numerical examples show that the pseudo-Newton method is up to 71 (in terms of CPU time) or 107 (in terms of number of iterations) times faster than the fastest QLB algorithm and thus makes bootstrap variance estimation feasible. Simulations and comparisons are performed and three real examples (Down syndrome data, kyphosis data, and colon microarray data) are analyzed to illustrate the proposed methods.
UR - http://www.scopus.com/inward/record.url?scp=40249087061&partnerID=8YFLogxK
U2 - 10.1016/j.csda.2007.11.007
DO - 10.1016/j.csda.2007.11.007
M3 - Journal article
AN - SCOPUS:40249087061
SN - 0167-9473
VL - 52
SP - 3528
EP - 3542
JO - Computational Statistics and Data Analysis
JF - Computational Statistics and Data Analysis
IS - 7
ER -