BriefGPT.xyz
Jun, 2020
深度神经网络的方向修剪
Directional Pruning of Deep Neural Networks
HTML
PDF
Shih-Kang Chao, Zhanyu Wang, Yue Xing, Guang Cheng
TL;DR
提出一种新的方向剪枝方法,用于在训练损失的平稳区域内或接近该区域内寻找稀疏解,证明了该方法在高度稀疏时对ResNet50,VGG16和wide ResNet 28x10等神经网络的同时达到与SGD相同的极小值,并且所找到的极小值不会影响训练损失
Abstract
In the light of the fact that the
stochastic gradient descent
(SGD) often finds a flat minimum valley in the
training loss
, we propose a novel directional
→