xSVM: Scalable Distributed Kernel Support Vector Machine Training
Publication type: Proceedings Article
Publication date: 2019-12-01
Abstract
Kernel Support Vector Machine (SVM) is a popular machine learning model for classification and regression. A significant challenge of large scale Kernel SVM is the size of the Gram matrix $(n \times n)$, which cannot be stored or processed efficiently when training data-set is large (e.g. n in the millions). This paper proposes a novel SVM training algorithm and its parallelization strategy that can efficiently train on data-sets with millions of samples on thousands of processors. It consists of an accurate, fast, and scalable low rank matrix approximation based on random projection, and a primal-dual interior point method to solve the approximated optimization problem. We demonstrate that xSVM is fast, scalable, and accurate on large scale data-sets and computing nodes. Compared to state-of-the-art distributed Kernel L1-SVM system xSVM is consistently several times faster, with comparable accuracy to the exact model trained by LIBSVM.
Found
Found
Top-30
Journals
1
|
|
Journal of Energy Chemistry
1 publication, 33.33%
|
|
1
|
Publishers
1
|
|
Institute of Electrical and Electronics Engineers (IEEE)
1 publication, 33.33%
|
|
Association for Computing Machinery (ACM)
1 publication, 33.33%
|
|
Elsevier
1 publication, 33.33%
|
|
1
|
- We do not take into account publications without a DOI.
- Statistics recalculated only for publications connected to researchers, organizations and labs registered on the platform.
- Statistics recalculated weekly.
Are you a researcher?
Create a profile to get free access to personal recommendations for colleagues and new articles.