Abstract

The dramatic evolution in the Deep Learning (DL) algorithms required to alter the silicon architecture fabric of the conventional parallel processing units to increase the efficiency of accelerationg the enormous feature data while achieving reasonable low power consumption levels, especially for the Convolutional Neural Networks (CNN). In this paper, three proposed Pyramidal Neuron Accelerator Architecture (PNAA) units have been designed and optimized for accelerating the convolutional layer of the Convolutional Neural Networks (CNN). The three proposed PNAA units are suggested to replace the conventional generic embedded Digital Signal Processing (DSP) blocks in the silicon architecture fabric of the FPGA chips, that are responsible for the dot-operation functions. The three proposed PNAA units represent the intensively-used neuron operations for the most common kernel filter dimensions in CNN systems as a proof of concept. The proposed PNAA units have been compiled using the TSMC 130 nm technology using the Synopsys DC compiler software. The Front-End analysis for different characteristic PVTs showed that the maximum achieved processing speed for the proposed PNAA units could reach a computational speed of 20.9 Giga Operation per Seconds (GPOS) at a frequency of 409.84 MHz and a predicted power consumption equal to 58.729 mW.

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.