K. Koukos A. Ros E. Hagersten

S Kaxiras, Z Wang, X Wang, F Hou, Y Luo,S Chai,M Isnardi,S Lim, R Karri,S Akram,JB Sartor,K Van Craeynest, W Heirman,L Eeckhout

ACM Transactions on(2016)

引用 0|浏览12
暂无评分
摘要
Memory access latency has become one of the major bottlenecks to system performance. To hide memory access latency, researchers have proposed hardware data prefetching mechanisms. Prefetching refers to fetching data from memory into caches or a prefetching buffer before actually using them. In this way, the memory latency can be effectively hidden and the processors performance can be improved. Data prefetching techniques in conventional single-core processors have been proven useful. However, on a chip multiprocessor (CMP) system, all running threads usually share the last level cache (LLC) and off-chip memory. Memory requests from each core conflict with those from other cores, since prefetching requests need to traverse on-chip interconnect networks and the memory bus to arrive at memory banks as normal data requests. The presence of a prefetching engine would cause additional cache …
更多
查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要