BriefGPT.xyz
Oct, 2017
剪枝还是不剪枝:探索模型压缩中剪枝的有效性
To prune, or not to prune: exploring the efficacy of pruning for model compression
HTML
PDF
Michael Zhu, Suyog Gupta
TL;DR
本文探讨在资源受限环境下,通过模型剪枝来压缩神经网络模型的方法,提出了一种简单、直接、易于应用的逐渐剪枝技术,并在多个模型/数据集上进行了比较,发现大型稀疏模型在保持较高精度的同时可减少10倍的参数数量。
Abstract
model pruning
seeks to induce sparsity in a
deep neural network
's various connection matrices, thereby reducing the number of nonzero-valued parameters in the model. Recent reports (Han et al., 2015; Narang et al
→