Publication | Closed Access
Fixed point optimization of deep convolutional neural networks for object recognition
242
Citations
12
References
2015
Year
Unknown Venue
Convolutional Neural NetworkEngineeringMachine LearningAutoencodersSpeech RecognitionImage ClassificationPattern RecognitionSparse Neural NetworkHigh PrecisionL2 Error MinimizationMachine VisionObject DetectionComputer EngineeringComputer ScienceDeep LearningNeural Architecture SearchModel CompressionComputer VisionDeep Neural NetworksObject RecognitionQuantized WeightsPoint OptimizationSpeech Processing
Deep convolutional neural networks have shown promising results in image and speech recognition applications. The learning capability of the network improves with increasing depth and size of each layer. However this capability comes at the cost of increased computational complexity. Thus reduction in hardware complexity and faster classification are highly desired. This work proposes an optimization method for fixed point deep convolutional neural networks. The parameters of a pre-trained high precision network are first directly quantized using L2 error minimization. We quantize each layer one by one, while other layers keep computation with high precision, to know the layer-wise sensitivity on word-length reduction. Then the network is retrained with quantized weights. Two examples on object recognition, MNIST and CIFAR-10, are presented. Our results indicate that quantization induces sparsity in the network which reduces the effective number of network parameters and improves generalization. This work reduces the required memory storage by a factor of 1/10 and achieves better classification results than the high precision networks.
| Year | Citations | |
|---|---|---|
Page 1
Page 1