BriefGPT.xyz
Oct, 2022
神经机器翻译中编码器-解码器是否冗余?
Is Encoder-Decoder Redundant for Neural Machine Translation?
HTML
PDF
Yingbo Gao, Christian Herold, Zijian Yang, Hermann Ney
TL;DR
本研究探讨了将源文本和目标文本直接拼接并训练语言模型进行翻译的想法,通过对双语翻译、额外目标语单语数据翻译和多语言翻译的实验,结果表明这种替代方法与基线中的编码器 - 解码器Transformer模型表现相当,表明编码器 - 解码器架构对于神经机器翻译可能是多余的。
Abstract
encoder-decoder architecture
is widely adopted for sequence-to-sequence modeling tasks. For
machine translation
, despite the evolution from long short-term memory networks to Transformer networks, plus the introd
→