Abstract
Binarized neural network (BNN) enables resistive switching random access memory (RRAM) with high nonlinearity and nonsymmetry to realize online training, using an RRAM & comparator structure. In this work, a new hardware implementation approach is proposed to improve the efficiency of BNN. In the approach, an 1T1R array-based propagation module is introduced and designed to realize the computing acceleration of fully parallel vector-matrix multiplication (VMM) in both forward and backward propagations. Using the 1T1R-based propagation module, high computing efficiency is achieved in both training and inference tasks, improving by 50× and 177×, respectively. To solve the computation error caused by device variation, a novel operation scheme with low gate voltage is proposed. With the operation scheme, the RRAM variation is dramatically suppressed by 74.8% for cycleto-cycle and 59.9% for device-to-device. It enables high-accuracy VMM calculation and, therefore, achieves 94.7% accuracywith a typical BNN, showing only 0.7% degradation from the ideal variation-free case.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.