谷歌浏览器插件
订阅小程序
在清言上使用

High-Throughput Multi-Threaded Sum-Product Network Inference in the Reconfigurable Cloud

International Workshop on Heterogeneous High-performance Reconfigurable Computing(2019)

引用 8|浏览13
暂无评分
摘要
Large cloud providers have started to make powerful FPGAs available as part of their public cloud offers. One promising application area for this kind of instances is the acceleration of machine learning tasks. This work presents an accelerator architecture that uses multiple accelerator cores for the inference in so-called Sum-Product Networks and complements it with a host software interface that overlaps data-transfer and actual computation. The evaluation shows that, the proposed architecture deployed to Amazon AWS F1 instances is able to outperform a 12-core Xeon processor by a factor of up to 1.9x and a Nvidia Tesla V100 GPU by a factor of up to 6.6x.
更多
查看译文
关键词
FPGA,SPN,Machine Learning,Graphical Models,Deep Models,Cloud
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要