Block-Wise Mixed-Precision Quantization: Enabling High Efficiency for Practical ReRAM-based DNN Accelerators
IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS(2024)
Key words
Quantization (signal),Computer architecture,Computational modeling,Artificial neural networks,Performance evaluation,Inference algorithms,Training,Deep neural network (DNN) acceleration,mixed-precision quantization,model compression,processing-in-memory (PIM),resistive random access memory (ReRAM)
AI Read Science
Must-Reading Tree
Example

Generate MRT to find the research sequence of this paper
Chat Paper
Summary is being generated by the instructions you defined