FPGA-based Accelerator for Long Short-Term Memory Recurrent Neural Networks

2017 22nd Asia and South Pacific Design Automation Conference (ASP-DAC)(2017)

引用 224|浏览171
暂无评分
摘要
Long Short-Term Memory Recurrent neural networks (LSTM-RNNs) have been widely used for speech recognition, machine translation, scene analysis, etc. Unfortunately, general-purpose processors like CPUs and GPGPUs can not implement LSTM-RNNs efficiently due to the recurrent nature of LSTM-RNNs. FPGA-based accelerators have attracted attention of researchers because of good performance, high energy-efficiency and great flexibility. In this work, we present an FPGA-based accelerator for LSTM-RNNs that optimizes both computation performance and communication requirements. The peak performance of our accelerator achieves 7.26 GFLOP/S, which significantly outperforms previous approaches.
更多
查看译文
关键词
FPGA-based accelerator,long short-term memory recurrent neural networks,LSTM-RNN,speech recognition,machine translation,scene analysis,computation performance,communication requirements
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要