BriefGPT.xyz
Apr, 2023
协调混乱: 神经网络训练的运行差异是无害且不可避免的
Calibrated Chaos: Variance Between Runs of Neural Network Training is Harmless and Inevitable
HTML
PDF
Keller Jordan
TL;DR
本文研究了神经网络训练过程中的表现差异性,找出了其来源,证明了其不可避免的存在性,并通过初步研究从各个方面探讨了训练初值、数据扩增和学习率等方面的影响。
Abstract
Typical
neural network trainings
have substantial
variance
in
test-set performance
between repeated runs, impeding
→