BriefGPT.xyz
Apr, 2020
关于低资源语言翻译的最优Transformer深度
On optimal transformer depth for low-resource language translation
HTML
PDF
Elan van Biljon, Arnu Pretorius, Julia Kreutzer
TL;DR
本研究对Transformer模型在低资源语言翻译中的应用进行了探究,发现过度追求模型大小存在负面影响,需要注意调整超参数以提高性能。同时,本研究旨在挖掘更佳的模型性能,以推动“Masakhane”项目的发展。
Abstract
transformers
have shown great promise as an approach to
neural machine translation
(NMT) for
low-resource languages
. However, at the same
→