Publication | Closed Access
Concurrent MAC unit design using VHDL for deep learning networks on FPGA
19
Citations
9
References
2018
Year
Unknown Venue
Unit ArchitectureEngineeringMachine LearningHardware AccelerationAdvanced ComputingHardware AlgorithmConvolutional Neural NetworksComputer EngineeringComputer ArchitectureFully-customize Mac UnitEmbedded Machine LearningParallel ProgrammingComputer ScienceParallel ComputingDeep LearningNeural Architecture SearchFpga DesignDeep Learning Networks
Deep neural network algorithms have proven their enormous capabilities in wide range of artificial intelligence applications, specially in Printed/Handwritten text recognition, Multimedia processing, Robotics and many other high end technological trends. The most challenging aspect nowadays is to overcome the extremely computational processing demands in applying such algorithms, especially in real-time systems. Recently, the Field Programmable Gate Array (FPGA) has been considered as one of the optimum hardware accelerator platform for accelerating the deep neural network architectures due to its large adaptability and the high degree of parallelism it offers. In this paper, the proposed 8-bits fixed-point parallel multiply-accumulate (MAC) unit architecture aimed to create a fully-customize MAC unit for the Convolutional Neural Networks (CNN) instead of depending on the conventional DSP blocks and embedded memories units on the FPGAs architecture silicon fabrics. The proposed 8-bits fixed-point parallel multiply-accumulate (MAC) unit architecture is designed using VHDL language and can performs a computational speed up to 4.17 Giga Operation per Second (GOPS) using high-density FPGAs.
| Year | Citations | |
|---|---|---|
Page 1
Page 1