Efficient Residue Number System Based Winograd Convolution

European Conference on Computer Vision(2020)

引用 18|浏览35
暂无评分
摘要
Prior research has shown that Winograd algorithm can reduce the computational complexity of convolutional neural networks (CNN) with weights and activations represented in floating point. However it is difficult to apply the scheme to the inference of low-precision quantized (e.g. INT8) networks. Our work extends the Winograd algorithm to Residue Number System (RNS). The minimal complexity convolution is computed precisely over large transformation tile (e.g. \(10 \times 10\) to \(16 \times 16\)) of filters and activation patches using the Winograd transformation and low cost (e.g. 8-bit) arithmetic without degrading the prediction accuracy of the networks during inference. The arithmetic complexity reduction is up to \(7.03\times \) while the performance improvement is up to \(2.30\times \) to \(4.69\times \) for \(3\times 3\) and \(5 \times 5\) filters respectively.
更多
查看译文
关键词
residue
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要