15.5 A 28nm 64kb 6t Sram Computing-In-Memory Macro With 8b Mac Operation For Ai Edge Chips

2020 IEEE INTERNATIONAL SOLID- STATE CIRCUITS CONFERENCE (ISSCC)(2020)

引用 163|浏览81
暂无评分
摘要
Advanced AI edge chips require multibit input (IN), weight (W), and output (OUT) for CNN multiply-and-accumulate (MAC) operations to achieve an inference accuracy that is sufficient for practical applications. Computing-in-memory (CIM) is an attractive approach to improve the energy efficiency $(\mathrm{EF}_{\mathrm{MAC}}]$ of MAC operations under a memory-wall constraint. Previous SRAM-CIM macros demonstrated a binary MAC [4], an in-array 8b W-merging with near-memory computing (NMC) using 6T SRAM cells (limited output precision) [5], a 7b1N-1 bW MAC using a 10T SRAM cell (large area) [3], an 4b1N-5bW MAC with a T8T SRAM cell [1], and 8b1N-1bW NMC with 8T SRAM (long MAC latency $(T_{\mathrm{AC}})$ ) [2]. However, previous works have not achieved high IN/W/OUT precision with fast $\mathrm{T}_{\mathrm{AC}}$ compact-area, high $\mathrm{EF}_{\mathrm{MAC}}$ , and robust readout against process variation, due to (1) small sensing margin in word-wise multiple-bit MAC operations, (2) a tradeoff between read accuracy vs. area overhead under process variation, (3) limited $\mathrm{EF}_{\mathrm{MAC}}$ due to decoupling of software and hardware development.
更多
查看译文
关键词
binary MAC,SRAM computing-in-memory macro,word-wise multiple-bit MAC operations,long MAC latency,8T SRAM,8b1N-1bW NMC,T8T SRAM cell,4b1N-5bW MAC,10T SRAM cell,7b1N-1 bW MAC,6T SRAM cells,near-memory computing,previous SRAM-CIM macros,memory-wall constraint,advanced AI edge chips,size 28.0 nm,word length 8 bit
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要