TL;DR通过将输入嵌入和输出投影矩阵进行绑定,利用更多信息和减小可训练变量数量,提出了一种提高自然语言模型学习效率的新型理论框架,并在Penn Tree Bank数据集上实现了最先进的性能。
Abstract
recurrent neural networks have been very successful at predicting sequences of words in tasks such as language modeling. However, all such models are based on the conventional classification framework, where mode