Concepedia

Publication | Closed Access

Energy-efficient, low-latency realization of neural networks through boolean logic minimization

37

Citations

8

References

2019

Year

Abstract

Deep neural networks have been successfully deployed in a wide variety of applications including computer vision and speech recognition. To cope with computational and storage complexity of these models, this paper presents a training method that enables a radically different approach for realization of deep neural networks through Boolean logic minimization. The aforementioned realization completely removes the energy-hungry step of accessing memory for obtaining model parameters, consumes about two orders of magnitude fewer computing resources compared to realizations that use floating-point operations, and has a substantially lower latency.

References

YearCitations

Page 1