Abstract

Binarized neural networks (BNNs) architecture play a vital role in the development of deep learning accelerator for memory-constrained IoT devices. However, the cost-efficiency of the domain-specific accelerators still requires a simplified structure to enhance the computing performance of BNNs. We propose and present a reconfigurable BNN accelerator to improve the computing speed through channel amplitude and an adaptive spatial amplitude model. To reduce the redundant operations in the binarization of matrix multiplication, we introduce the channel amplitude model for BNN convolution functions. The adaptive spatial amplitude is implemented with the help of a matrix multiplication layer with 3 × 3 convolutions to get spatial information and increases the computation speed. For hardware implementation, the channel amplitude conversion is deployed through XNOR-popcount convolutions of each layer. The performance results of our proposed accelerator provide 2–10 × line buffer efficiency, 1.6–3.7 × processing frequency enhancement, and 2.35 × processing element reduction compared with existing baselines using VGG-16 based FPGA implementation.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call