Chrome Extension
WeChat Mini Program
Use on ChatGLM

MemoryFormer: Minimize Transformer Computation by Removing Fully-Connected Layers

Ning Ding,Yehui Tang,Haochen Qin, Zhenli Zhou,Chao Xu, Lin Li,Kai Han, Liao Heng,Yunhe Wang

NeurIPS 2024(2024)

Cited 0|Views9
Key words
transformer architecture,large language model,computational complexity reduction
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined