谷歌浏览器插件
订阅小程序
在清言上使用

52.5 TOPS/W 1.7GHz Reconfigurable XGBoost Inference Accelerator Based on Modular-Unit-Tree with Dynamic Data and Compute Gating.

Chang Eun Song,Yidong Li, Amardeep Ramnani, Pulkit Agrawal, Purvi Agrawal,Sung-Joon Jang,Sang-Seol Lee,Tajana Rosing,Mingu Kang

2024 IEEE Custom Integrated Circuits Conference (CICC)(2024)

引用 0|浏览4
暂无评分
摘要
The XGBoost has emerged as a powerful AI algorithm achieving high accuracy winning multiple Kaggle competitions in many tasks including medical diagnosis, recommendation system, and autonomous driving [1]. It provides a great potential for edge devices due to a binary-tree-based simple computing kernel compared to deep learning [2]. Despite such a potential from the kernel-level simplicity, the efficient end-to-end realization is hindered by multiple design challenges due to 1) the highly irregular tree shape, 2) low hardware utilization, 3) delay from the sequential processing of each tree node, and 4) a large data movement to all nodes [3]–[5]. We propose low-power and high performance XGBoost accelerator by employing modular unit trees and reconfigurable interconnects along with a selective data movement and execution. The proposed accelerator achieves 52.5 TOPS/Wand 0.41 $\text{TOPS}/\text{mm}^{2}$ , which are the best among the reported CNN and tree-based classifiers [2]–[5].
更多
查看译文
关键词
XGBoost,Input Features,Light Signal,Lookup Table,Load Data,Leaf Node,Recommender Systems,Multiple Trees,High Energy Cost,Tree Shape,Modular Units,Tree-based Classifiers,Tree Architecture,Kernel Computation,Output Gain,CNN Classifier
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要