Impacts and solutions of nonvolatile-memory-induced weight error in the computing-in-memory neural network system

JAPANESE JOURNAL OF APPLIED PHYSICS(2020)

引用 1|浏览71
暂无评分
摘要
Nonvolatile-memory-based computing-in-memory architecture is one of the solutions to the massive data movement problem in the conventional von Neumann computing architecture since multiplication-and-accumulation (MAC) operations can be directly performed inside the memory array. This paper investigates the errors from the imperfections of resistive random access memory, including program error, read fluctuation and retention drift, and their impacts on the inference accuracy in convolutional neural network. The influences from weight errors in each convolution layer are evaluated according to the change of neuron distributions. A batch normalization (BN) parameter calibration method is proposed in order to correctly scale-and-shift the MAC results to compensate weight errors. This calibrated BN process drastically improves the inference accuracy not only for as-programmed analog ReRAM array but also for devices after longtime retention. This approach provides an effective direction to deal with the nonvolatile-memory-induced errors in artificial neural networks. (C) 2020 The Japan Society of Applied Physics
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要