Z-PIM: An Energy-Efficient Sparsity Aware Processing-In-Memory Architecture with Fully-Variable Weight Precision

2020 IEEE Symposium on VLSI Circuits(2020)

引用 13|浏览39
暂无评分
摘要
This paper presents Z-PIM, an energy-efficient processing-in-memory (PIM) architecture that supports zero-skipping operations and fully-variable weight bit-precision for efficient deep neural network (DNN). The 8T-SRAM cell based bit-serial operation with hierarchical bit-line structure enables variable weight precision and reduces bit-line switching by 95.42% in convolution layers of VGG-16. Z-PIM handles abundant zeros in weight data by skip-reading their corresponding input data while read-sequence rearranging and pipelining improves throughput by 66.1%. In addition, diagonal accumulation logic is proposed to accumulate both partial-sums for bit-serial operation and spatial products. As a result, the Z-PIM chip fabricated in a 65nm process consumes average 5.294mW power and achieves 0.31-49.12 TOPS/W energy efficiency for convolution operations as sparsity and weight bit-precision vary from 0.1 to 0.9 and 1b to 16b, respectively.
更多
查看译文
关键词
energy-efficient processing-in-memory architecture,fully-variable weight bit-precision,deep neural network,8T-SRAM cell based bit-serial operation,hierarchical bit-line structure,bit-line switching,read-sequence rearranging,Z-PIM chip,convolution operations,energy-efficient sparsity aware processing-in-memory architecture,zero-skipping operations,diagonal accumulation logic,size 65.0 nm,power 5.294 mW,word length 1 bit to 16 bit
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要