谷歌浏览器插件
订阅小程序
在清言上使用

Dependent scalar quantization for neural network compression

2020 IEEE International Conference on Image Processing (ICIP)(2020)

引用 8|浏览24
暂无评分
摘要
Recent approaches to compression of deep neural networks, like the emerging standard on compression of neural networks for multimedia content description and analysis (MPEG-7 part 17), apply scalar quantization and entropy coding of the quantization indexes. In this paper we present an advanced method for quantization of neural network parameters, which applies dependent scalar quantization (DQ) or trellis-coded quantization (TCQ), and an improved context modeling for the entropy coding of the quantization indexes. We show that the proposed method achieves 5.778% bitrate reduction and virtually no loss (0.37%) of network performance in average, compared to the baseline methods of the second test model (NCTM) of MPEG-7 part 17 for relevant working points.
更多
查看译文
关键词
dependent scalar quantization,trellis-coded quantization,entropy coding,arithmetic coding,neural network compression
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要