Publication | Closed Access
TSUNAMI: Triple Sparsity-Aware Ultra Energy-Efficient Neural Network Training Accelerator With Multi-Modal Iterative Pruning
16
Citations
25
References
2022
Year
EngineeringMachine LearningHardware AccelerationEnergy EfficiencySparse Neural NetworkComputer EngineeringComputer ArchitectureDomain-specific AcceleratorEmbedded Machine LearningComputer ScienceParallel ComputingDeep LearningNeural Architecture SearchWeight Sparsity ImbalanceModel CompressionWeight MemoryMulti-modal Iterative Pruning
This article proposes the TSUNAMI, which supports an energy-efficient deep-neural-network training. The TSUNAMI supports multi-modal iterative pruning to generate zeros in activation and weight. Tile-based dynamic activation pruning unit and weight memory shared pruning unit eliminate additional memory access. Coarse-zero skipping controller skips multiple unnecessary multiply-and-accumulation (MAC) operations at once, and fine-zero skipping controller skips randomly located unnecessary MAC operations. Weight sparsity balancer solves a utilization degradation caused by weight sparsity imbalance, and the workload of each convolution core is allocated by a random channel allocator. The TSUNAMI achieves an energy efficiency of 3.42 TFLOPS/W at 0.78V and 50MHz with floating-point 8-bit activation and weight. Also, it achieves an energy efficiency of 405.96 TFLOPS/W at 90% sparsity condition.
| Year | Citations | |
|---|---|---|
Page 1
Page 1