BriefGPT.xyz
Jun, 2017
基于注意力机制的NMT词汇选择
Attention-based Vocabulary Selection for NMT Decoding
HTML
PDF
Baskaran Sankaran, Markus Freitag, Yaser Al-Onaizan
TL;DR
提出一种使用 NMT 训练过程中直接从 attention layer 中学习候选列表的新方法,该方法高度优化当前 NMT 模型的候选列表,无须对候选池进行外部计算并实现了在不降低翻译质量的情况下显著提高解码速度。
Abstract
neural machine translation
(NMT) models usually use large
target vocabulary
sizes to capture most of the words in the target language. The vocabulary size is a big factor when decoding new sentences as the final
→