Minimalist Design for Accelerating Convolutional Neural Networks for Low-End FPGA Platforms
2017 IEEE 25th Annual International Symposium on Field-Programmable Custom Computing Machines (FCCM)(2017)
摘要
Deep neural networks have gained tremendous attention in both the academic and industrial communities due to their performance in many artificial intelligence applications, particularly in computer vision. However, these algorithms are known to be computationally very demanding for both scoring and model learning applications. State-of-the-art recognition models use tens of millions of parameters and have significant memory and computational requirements. These requirements have restricted the users of deep neural network applications to high-end, expensive, and power hungry IoT platforms to penetrate the deep learning markets. This paper presents work at the leading edge intersection of several evolving technologies, including emerging IoT platforms, Deep Learning, and Field-programmable Gate Array (FPGA) computing. We demonstrate a new minimalist design methodology that minimizes the utilization of FPGA resources and can run deep learning algorithms with over 60 million parameters. This makes particularly suitable for resource-constrained, low-end FPGA platforms.
更多查看译文
关键词
minimalist design,convolutional neural networks,low-end FPGA platforms,deep neural networks,high-end IoT platforms,deep learning,field-programmable gate array,FPGA computing,FPGA resource utilization,resource-constrained FPGA platforms
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络