VTA: An Open Hardware-Software Stack for Deep Learning.

arXiv: Learning(2018)

引用 78|浏览337
暂无评分
摘要
Hardware acceleration is an enabler for ubiquitous and efficient deep learning. With hardware accelerators being introduced in datacenter and edge devices, it is time to acknowledge that hardware specialization is central to the deep learning system stack. technical report presents the Versatile Tensor Accelerator (VTA), an open, generic, and customizable deep learning accelerator design. is a programmable accelerator that exposes a RISC-like programming abstraction to describe operations at the tensor level. We designed to expose the most salient and common characteristics of mainstream deep learning accelerators, such as tensor operations, DMA load/stores, and explicit compute/memory arbitration. VTA is more than a standalone accelerator design: itu0027s an end-to-end solution that includes drivers, a JIT runtime, and an optimizing compiler stack based on TVM. The current release of includes a behavioral hardware simulator, as well as the infrastructure to deploy on low-cost FPGA development boards for fast prototyping. By extending the TVM stack with a customizable, and open source deep learning hardware accelerator design, we are exposing a transparent end-to-end deep learning stack from the high-level deep learning framework, down to the actual hardware design and implementation. This forms a truly end-to-end, from software-to-hardware open source stack for deep learning systems.
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要