BriefGPT.xyz
Jun, 2020
神经网络中简化偏差的陷阱
The Pitfalls of Simplicity Bias in Neural Networks
HTML
PDF
Harshay Shah, Kaustav Tamuly, Aditi Raghunathan, Prateek Jain, Praneeth Netrapalli
TL;DR
本文旨在通过设计包含不同时简单性的多个预测特征的数据集,捕捉实际训练数据中的非鲁棒性,从理论和实证研究中发现简洁性偏见在训练神经网络中的作用及其对泛化和鲁棒性的影响,提出新算法以避免简洁性偏见的缺陷。
Abstract
Several works have proposed
simplicity bias
(SB)---the tendency of standard training procedures such as Stochastic Gradient Descent (SGD) to find simple models---to justify why
neural networks
generalize well [Ar
→