xSVM: Scalable Distributed Kernel Support Vector Machine Training

Publication typeProceedings Article
Publication date2019-12-01
Abstract
Kernel Support Vector Machine (SVM) is a popular machine learning model for classification and regression. A significant challenge of large scale Kernel SVM is the size of the Gram matrix $(n \times n)$, which cannot be stored or processed efficiently when training data-set is large (e.g. n in the millions). This paper proposes a novel SVM training algorithm and its parallelization strategy that can efficiently train on data-sets with millions of samples on thousands of processors. It consists of an accurate, fast, and scalable low rank matrix approximation based on random projection, and a primal-dual interior point method to solve the approximated optimization problem. We demonstrate that xSVM is fast, scalable, and accurate on large scale data-sets and computing nodes. Compared to state-of-the-art distributed Kernel L1-SVM system xSVM is consistently several times faster, with comparable accuracy to the exact model trained by LIBSVM.
Found 
Found 

Top-30

Journals

1
Journal of Energy Chemistry
1 publication, 33.33%
1

Publishers

1
Institute of Electrical and Electronics Engineers (IEEE)
1 publication, 33.33%
Association for Computing Machinery (ACM)
1 publication, 33.33%
Elsevier
1 publication, 33.33%
1
  • We do not take into account publications without a DOI.
  • Statistics recalculated only for publications connected to researchers, organizations and labs registered on the platform.
  • Statistics recalculated weekly.

Are you a researcher?

Create a profile to get free access to personal recommendations for colleagues and new articles.
Metrics
Share
Found error?