Abstract
In order to explore the efficient design and implementation of activation function in deep learning processor, this paper presents an efficient five-stage pipelined hardware architecture for activation function based on the piecewise linear interpolation, and a novel neuron data-LUT address mapping algorithm. Compared with the previous designs based on serial calculation, the proposed hardware architecture can achieve at least 3 times of acceleration. Four commonly used activation functions are designed based on the proposed hardware architecture, which is implemented on the XC6VLX240T of Xilinx. The LeNet-5 and AlexNet are selected as benchmarks to test the inference accuracy of different activation functions with different piecewise numbers on the MNIST and CIFAR-10 test sets in the deep learning processor prototype system. The experiment results show that the proposed hardware architecture can effectively accomplish the relevant calculation of activation functions in the deep learning processor and the accuracy loss is negligible. The proposed hardware architecture is adaptable for numerous activation functions, which can be widely used in the design of other deep learning processors.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.