Abstract

A deep neural network (DNN) learning processor, DF-LNPU, is proposed for fast online learning by utilizing direct feedback alignment (DFA). The proposed processor develops the pipelined DFA (PDFA) and shows 1.75–3.05 $\times $ faster DNN learning speed compared with the previous processors. The processor consists of two different types of core. Due to the heterogeneous characteristics, the DF-LNPU shows higher area and energy efficiency than the homogeneous approach. Besides, inter-core and intra-core pipeline is constructed for high-speed online learning and reduces the overall processing time by 42.8–68.4%. Finally, the direct error propagation core (DEPC) is proposed with the built-in pseudorandom number generator (PRNG). The DEPC adopts binarized DFA and organizes the adder-only computing units to maximize computation efficiency. The PRNG-based backward weight generation reduces overall external memory access by 42.8%, and adder-only error computation improves the area and energy efficiency by 35.0% and 14.3%, respectively. The DF-LNPU is implemented in 65-nm CMOS technology, and it can be operated from 0.7-to-1.1-V supply voltage with a maximum frequency of 200 MHz. PDFA-based computing enables 80.94-OP/KB throughput per memory bandwidth, which is the best figure compared with the back-propagation dependent learning processors. The functionality of the DF-LNPU was successfully demonstrated on the verification system using an MDNet-based object tracking application.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call