High Performance Training Of Deep Neural Networks Using Pipelined Hardware Acceleration And Distributed Memory
2018 19TH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED)(2018)
摘要
Recently, Deep Neural Networks (DNNs) have made unprecedented progress in various tasks. However, there is a timely need to accelerate the training process in DNNs specifically for real-time applications that demand high performance, energy efficiency and compactness. Numerous algorithms have been proposed to improve the accuracy, however the network training process is computationally slow. In this paper, we present a scalable pipelined hardware architecture with distributed memories for a digital neuron to implement deep neural networks. We also explore various functions and algorithms as well as different memory topologies, to optimize the performance of our training architecture. The power, area, and delay of our proposed model are evaluated with respect to software implementation. Experimental results on the MNIST dataset demonstrate that compared with the software training, our proposed hardware based approach for training process achieves 33X runtime reduction, 5X power reduction, and nearly 168X energy reduction.
更多查看译文
关键词
Deep learning, neural network, hardware design
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络