Publication | Closed Access
CA-SVM: Communication-Avoiding Support Vector Machines on Distributed Systems
38
Citations
23
References
2015
Year
Unknown Venue
Cluster ComputingEngineeringMachine LearningDistributed AlgorithmsProblem SizeComputational ComplexityParallel AlgorithmsDistributed Memory ClustersSupport Vector MachineData ScienceData MiningPattern RecognitionComputing SystemsStatistical Machine LearningParallel ComputingDistributed ModelKnowledge DiscoveryDistributed SystemsComputer ScienceSignal ProcessingParallel ProcessingParallel LearningParallel ProgrammingData-level ParallelismKernel MethodVectorization
We consider the problem of how to design and implement communication-efficient versions of parallel support vector machines, a widely used classifier in statistical machine learning, for distributed memory clusters and supercomputers. The main computational bottleneck is the training phase, in which a statistical model is built from an input data set. Prior to our study, the parallel is efficiency of a state-of-the-art implementation scaled as W = Omega(P <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">3</sup> ), where W is the problem size and P the number of processors, this scaling is worse than even a one-dimensional block row dense matrix vector multiplication, which has W = Omega(P <sup xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">2</sup> ). This study considers a series of algorithmic refinements, leading ultimately to a Communication-Avoiding SVM (CASVM) method that improves the is efficiency to nearly W = Omega(P). We evaluate these methods on 96 to 1536 processors, and show average speedups of 3 - 16× (7× on average) over Dis-SMO, and a 95% weak-scaling efficiency on six real world datasets, with only modest losses in overall classification accuracy. The source code can be downloaded at https://github.com/fastalgo/casvm.
| Year | Citations | |
|---|---|---|
Page 1
Page 1