Publication | Closed Access
Training Sparse Neural Networks
170
Citations
24
References
2017
Year
Unknown Venue
Convolutional Neural NetworkEngineeringMachine LearningAutoencodersImage AnalysisData SciencePattern RecognitionSparse Neural NetworkSparse Neural NetworksSparse ComputationsSupervised LearningMachine VisionFeature LearningComputer ScienceNeural NetworksDeep LearningComputer VisionDeep Neural NetworksSparse Representation
The emergence of Deep neural networks has seen human-level performance on large scale computer vision tasks such as image classification. However these deep networks typically contain large amount of parameters due to dense matrix multiplications and convolutions. As a result, these architectures are highly memory intensive, making them less suitable for embedded vision applications. Sparse Computations are known to be much more memory efficient. In this work, we train and build neural networks which implicitly use sparse computations. We introduce additional gate variables to perform parameter selection and show that this is equivalent to using a spike-and-slab prior. We experimentally validate our method on both small and large networks which result in highly sparse neural network models.
| Year | Citations | |
|---|---|---|
Page 1
Page 1