Sciweavers

ECML
2007
Springer

Fast Optimization Methods for L1 Regularization: A Comparative Study and Two New Approaches

13 years 10 months ago
Fast Optimization Methods for L1 Regularization: A Comparative Study and Two New Approaches
L1 regularization is effective for feature selection, but the resulting optimization is challenging due to the non-differentiability of the 1-norm. In this paper we compare state-of-the-art optimization techniques to solve this problem across several loss functions. Furthermore, we propose two new techniques. The first is based on a smooth (differentiable) convex approximation for the L1 regularizer that does not depend on any assumptions about the loss function used. The other technique is a new strategy that addresses the non-differentiability of the L1-regularizer by casting the problem as a constrained optimization problem that is then solved using a specialized gradient projection method. Extensive comparisons show that our newly proposed approaches consistently rank among the best in terms of convergence speed and efficiency by measuring the number of function evaluations required.
Mark Schmidt, Glenn Fung, Rómer Rosales
Added 07 Jun 2010
Updated 07 Jun 2010
Type Conference
Year 2007
Where ECML
Authors Mark Schmidt, Glenn Fung, Rómer Rosales
Comments (0)