Share this page:

Coordinate Descent Method for Large-scale L2-loss Linear SVM

Kai-Wei Chang, Cho-Jui Hsieh, and Chih-Jen Lin, in JMLR, 2008.

Code

Download the full text


Abstract

Linear support vector machines (SVM) are useful for classifying large-scale sparse data. Problems with sparse features are common in applications such as document classification and natural language processing. In this paper, we propose a novel coordinate descent algorithm for training linear SVM with the L2-loss function. At each step, the proposed method minimizes a one-variable sub-problem while fixing other variables. The sub-problem is solved by Newton steps with the line search technique. The procedure globally converges at the linear rate. As each sub-problem involves only values of a corresponding feature, the proposed approach is suitable when accessing a feature is more convenient than accessing an instance. Experiments show that our method is more efficient and stable than state of the art methods such as Pegasos and TRON.


Bib Entry

@inproceedings{ChangHsLi08,
  author = {Chang, Kai-Wei and Hsieh, Cho-Jui and Lin, Chih-Jen},
  title = {Coordinate Descent Method for Large-scale L2-loss Linear SVM},
  booktitle = {JMLR},
  year = {2008}
}

Related Publications

  1. Large Linear Classification When Data Cannot Fit In Memory, TKDD, 2012
  2. Selective Block Minimization for Faster Convergence of Limited Memory Large-scale Linear Models, KDD, 2011
  3. Iterative Scaling and Coordinate Descent Methods for Maximum Entropy Models, JMLR, 2010
  4. A Comparison of Optimization Methods and software for Large-scale L1-regularized Linear Classification, JMLR, 2010
  5. Training and Testing Low-degree Polynomial Data Mappings via Linear SVM, JMLR, 2010
  6. A Sequential Dual Method for Large Scale Multi-Class Linear SVMs, KDD, 2008
  7. A Dual Coordinate Descent Method for Large-Scale Linear SVM, ICML, 2008
  8. LIBLINEAR: A Library for Large Linear Classification, JMLR, 2008