Abstract

Advanced AI edge chips require multibit input (IN), weight (W), and output (OUT) for CNN multiply-and-accumulate (MAC) operations to achieve an inference accuracy that is sufficient for practical applications. Computing-in-memory (CIM) is an attractive approach to improve the energy efficiency $(\mathrm{EF}_{\mathrm{MAC}}]$ of MAC operations under a memory-wall constraint. Previous SRAM-CIM macros demonstrated a binary MAC [4], an in-array 8b W-merging with near-memory computing (NMC) using 6T SRAM cells (limited output precision) [5], a 7b1N-1 bW MAC using a 10T SRAM cell (large area) [3], an 4b1N-5bW MAC with a T8T SRAM cell [1], and 8b1N-1bW NMC with 8T SRAM (long MAC latency $(T_{\mathrm{AC}})$ ) [2]. However, previous works have not achieved high IN/W/OUT precision with fast $\mathrm{T}_{\mathrm{AC}}$ compact-area, high $\mathrm{EF}_{\mathrm{MAC}}$ , and robust readout against process variation, due to (1) small sensing margin in word-wise multiple-bit MAC operations, (2) a tradeoff between read accuracy vs. area overhead under process variation, (3) limited $\mathrm{EF}_{\mathrm{MAC}}$ due to decoupling of software and hardware development.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call