BriefGPT.xyz
Jun, 2024
循环神经网络中的几何稀疏化
Geometric sparsification in recurrent neural networks
HTML
PDF
Wyatt Mackey, Ioannis Schizas, Jared Deighton, David L. Boothe, Jr., Vasileios Maroulas
TL;DR
稀疏化技术在大规模神经模型运行中减少计算成本的常用方法之一。本文提出了一种新的循环神经网络(RNNs)稀疏化技术,称为模量规则化,结合幅值修剪。通过使我们的规则化术语明确成几何形式,我们首次对我们的神经网络的期望稀疏架构进行了先验描述。验证了我们的方案对导航和自然语言处理RNNs的有效性。
Abstract
A common technique for ameliorating the computational costs of running large
neural models
is
sparsification
, or the removal of neural connections during training. Sparse models are capable of maintaining the hig
→