BriefGPT.xyz
Feb, 2018
批量卡尔曼归一化:向使用微批次训练深度神经网络迈进
Batch Kalman Normalization: Towards Training Deep Neural Networks with Micro-Batches
HTML
PDF
Guangrun Wang, Jiefeng Peng, Ping Luo, Xinjiang Wang, Liang Lin
TL;DR
本文提出了一种称为Batch Kalman Normalization(BKN)的新型规范化方法,通过考虑其前面所有层的分布来估计某个层的统计信息来改善和加速训练深度神经网络,特别是在微批处理的情况下,实现了更稳定的训练和更快的收敛速度。
Abstract
As an indispensable component,
batch normalization
(BN) has successfully improved the training of
deep neural networks
(DNNs) with mini-batches, by normalizing the distribution of the internal representation for
→