BriefGPT.xyz
Dec, 2015
具有注意力机制的前馈网络可以解决一些长期记忆问题
Feed-Forward Networks with Attention Can Solve Some Long-Term Memory Problems
HTML
PDF
Colin Raffel, Daniel P. W. Ellis
TL;DR
我们提出了一种适用于前馈神经网络的简化注意力模型,并证明其可解决比这些任务的最佳出版结果更长和更广泛变化的序列长度的综合“加法”和“乘法”长期记忆问题。
Abstract
Recurrent
neural networks
(RNNs) have proven to be powerful models in problems involving sequential data. Recently, RNNs have been augmented with "
attention
" mechanisms which allow the network to focus on differe
→