Abstract

Deep neural networks (DNNs) have been recently achieving state-of-the-art performance for many artificial intelligence (AI) applications such as computer vision, image recognition, and machine translator. Among them, image recognition using convolutional neural networks (CNNs) is widely used, but the implementation of CNN accelerator for mobile devices is largely restricted due to its intensive computation complexity and a large amount of memory access. In this paper, we adopt the heterogeneous SRAM sizing approach for the memories in CNN processor, where more important higher order data bits are stored in the relatively larger SRAM bit-cells and the less important bits are stored in the smaller ones. Numerical results with 65 nm technology show that compared to the conventional SRAM sizing, approximately 2% better accuracy in AlexNet is achieved using heterogeneous SRAM sizing under 500mV of supply voltage.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call