BriefGPT.xyz
Jul, 2016
层归一化
Layer Normalization
HTML
PDF
Jimmy Lei Ba, Jamie Ryan Kiros, Geoffrey E. Hinton
TL;DR
本文提出一种基于层归一化的深度神经网络训练新方法,能够有效稳定循环神经网络中的隐藏状态动态,其训练时间较之前的技术有大幅度降低。
Abstract
Training state-of-the-art,
deep neural networks
is computationally expensive. One way to reduce the
training time
is to normalize the activities of the neurons. A recently introduced technique called
→