Deep Neural Networks Pruning via the Structured Perspective Regularization

SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE(2023)

引用 0|浏览2
暂无评分
摘要
In machine learning, artificial neural networks (ANNs) are a very powerful tool, broadly used in many applications. Often, the selected (deep) architectures include many layers, and therefore a large number of parameters, which makes training, storage, and inference expensive. This motivated a stream of research about compressing the original networks into smaller ones without excessively sacrificing performances. Among the many proposed compression approaches, one of the most popular is pruning, whereby entire elements of the ANN (links, nodes, channels,...) and the corresponding weights are deleted. Since the nature of the problem is inherently combinatorial (what elements to prune and what not), we propose a new pruning method based on operational research tools. We start from a natural mixed-integer-programming model for the problem, and we use the perspective reformulation technique to strengthen its continuous relaxation. Projecting away the indicator variables from this reformulation yields a new regularization term, which we call the structured perspective regularization, that leads to structured pruning of the initial architecture. We test our method on some ResNet architectures applied to CIFAR-10, CIFAR-100, and ImageNet datasets, obtaining competitive performances w.r.t. the state of the art for structured pruning.
更多
查看译文
关键词
pruning,artificial neural networks,optimization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要