Abstract
The Orthant-Wise Limited memory QuasiNewton (OWL-QN) method has been demonstrated to be very effective in solving the regularized sparse learning problem. OWL-QN extends the L-BFGS from solving unconstrained smooth optimization problems to -regularized (non-smooth) sparse learning problems. At each iteration, OWL-QN does not involve any regularized quadratic optimization subproblem and only requires matrix-vector multiplications without an explicit use of the (inverse) Hessian matrix, which enables OWL-QN to tackle large-scale problems efficiently. Although many empirical studies have demonstrated that OWL-QN works quite well in practice, several recent papers point out that the existing convergence proof of OWL-QN is flawed and a rigorous convergence analysis for OWL-QN still remains to be established. In this paper, we propose a modified Orthant-Wise Limited memory Quasi-Newton (mOWL-QN) algorithm by slightly modifying the OWL-QN algorithm. As the main technical contribution of this paper, we establish a rigorous convergence proof for the mOWL-QN algorithm. To the best of our knowledge, our work fills the theoretical gap by providing the first rigorous convergence proof for the OWL-QN-type algorithm on solving ?1 regularized sparse learning problems. We also provide empirical studies to show that mOWLQN works well and is as efficient as OWL-QN. Proceedings of the 32 nd International Conference on MachLearning, Lille, France, 2015. JMLR: W&CP volume 37. Copyright 2015 by the author(s).