Publication | Closed Access
Channel Pruning for Accelerating Very Deep Neural Networks
2.5K
Citations
49
References
2017
Year
Unknown Venue
Convolutional Neural NetworkDeep Neural NetworksEngineeringMachine LearningPruned Vgg-16Sparse Neural NetworkNew ChannelComputer EngineeringComputer ScienceLasso RegressionDeep LearningNeural Architecture SearchModel Compression
In this paper, we introduce a new channel pruning method to accelerate very deep convolutional neural networks. Given a trained CNN model, we propose an iterative two-step algorithm to effectively prune each layer, by a LASSO regression based channel selection and least square reconstruction. We further generalize this algorithm to multi-layer and multi-branch cases. Our method reduces the accumulated error and enhance the compatibility with various architectures. Our pruned VGG-16 achieves the state-of-the-art results by 5× speed-up along with only 0.3% increase of error. More importantly, our method is able to accelerate modern networks like ResNet, Xception and suffers only 1.4%, 1.0% accuracy loss under 2× speedup respectively, which is significant.
| Year | Citations | |
|---|---|---|
2016 | 214.9K | |
2017 | 75.5K | |
2014 | 75.4K | |
2012 | 63.3K | |
2009 | 60.2K | |
1998 | 56.5K | |
2016 | 52.4K | |
1996 | 50.3K | |
2015 | 46.2K | |
2015 | 27.2K |
Page 1
Page 1