Chrome Extension
WeChat Mini Program
Use on ChatGLM

Smart-Infinity: Fast Large Language Model Training Using Near-Storage Processing on a Real System

2024 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE, HPCA 2024(2024)

Cited 0|Views129
No score
Key words
Processing in-memory/near-memory/in-cache,FPGA: Architectures and accelerators,Large Language Models (LLMs)
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined