Publication | Closed Access
A parallel computing platform for training large scale neural networks
32
Citations
16
References
2013
Year
Unknown Venue
Cluster ComputingEngineeringMachine LearningComputer ArchitectureParallel ImplementationLarge Scale TrainingMap-reduceData ScienceEmbedded Machine LearningParallel ComputingParallel Computing PlatformMassively-parallel ComputingComputer EngineeringComputer ScienceDeep LearningData-intensive ComputingLarge Scale AnnsArtificial Neural NetworksParallel ProcessingParallel LearningParallel ProgrammingData-level ParallelismMassive Data ProcessingBig Data
Artificial neural networks (ANNs) have been proved to be successfully used in a variety of pattern recognition and data mining applications. However, training ANNs on large scale datasets are both data-intensive and computation-intensive. Therefore, large scale ANNs are used with reservation for their time-consuming training to get high precision. In this paper, we present cNeural, a customized parallel computing platform to accelerate training large scale neural networks with the backpropagation algorithm. Unlike many existing parallel neural network training systems working on thousands of training samples, cNeural is designed for fast training large scale datasets with millions of training samples. To achieve this goal, firstly, cNeural adopts HBase for large scale training dataset storage and parallel loading. Secondly, it provides a parallel in-memory computing framework for fast iterative training. Third, we choose a compact, event-driven messaging communication model instead of the heartbeat polling model for instant messaging delivery. Experimental results show that the overhead time cost by data loading and messaging communication is very low in cNeural and cNeural is around 50 times faster than the solution based on Hadoop MapReduce. It also achieves nearly linear scalability and excellent load balancing.
| Year | Citations | |
|---|---|---|
Page 1
Page 1