BriefGPT.xyz
Sep, 2017
基于序列长度的线性循环神经网络并行化
Parallelizing Linear Recurrent Neural Nets Over Sequence Length
HTML
PDF
Eric Martin, Chris Cundy
TL;DR
该研究论文描述了使用线性顺序依赖关系的RNN可以使用并行扫描算法在序列长度上进行并行化训练,通过开发并行线性递归CUDA内核,加速多种最先进的RNN架构的训练和推理,扩展序列学习到以前无法触及的极长序列区域并成功训练GILR-LSTM进行一百万时间步长的合成序列分类任务。
Abstract
recurrent neural networks
(RNNs) are widely used to model sequential data but their non-linear dependencies between sequence elements prevent parallelizing training over sequence length. We show the training of RNNs with only
→