Distributed Block-diagonal Approximation Methods for Regularized Empirical Risk Minimization
Ching-pei Lee and Kai-Wei Chang, in Machine Learning Journal, 2019.
CodeDownload the full text
Abstract
Designing distributed algorithms for empirical risk minimization (ERM) has become an active research topic in recent years because of the practical need to deal with the huge volume of data. In this paper, we propose a general framework for training an ERM model via solving its dual problem in parallel over multiple machines. Our method provides a versatile approach for many large-scale machine learning problems, including linear binary/multi-class classification, regression, and structured prediction. Comparing with existing approaches, we show that our method has faster convergence under weaker conditions both theoretically and empirically.
Bib Entry
@inproceedings{LD17,
author = {Lee, Ching-pei and Chang, Kai-Wei},
title = {Distributed Block-diagonal Approximation Methods for Regularized Empirical Risk Minimization},
booktitle = {Machine Learning Journal},
year = {2019}
}
Related Publications
- Robust Text Classifier on Test-Time Budgets, EMNLP (short), 2019
- Efficient Contextual Representation Learning With Continuous Outputs, TACL, 2019
- Structured Prediction with Test-time Budget Constraints, AAAI, 2017
- A Credit Assignment Compiler for Joint Prediction, NeurIPS, 2016
- Learning to Search for Dependencies, Arxiv, 2015
- Learning to Search Better Than Your Teacher, ICML, 2015
- Structural Learning with Amortized Inference, AAAI, 2015