BriefGPT.xyz
Jun, 2020
神经序列建模任务损失最小化参数搜索中的MLE指导
MLE-guided parameter search for task loss minimization in neural sequence modeling
HTML
PDF
Sean Welleck, Kyunghyun Cho
TL;DR
本文提出了一种名为MGS的新方法,其中基于随机搜索参数空间的分布,使用非确定性方法引导参数更新方向,从而优化序列级别的任务损失,实现了显著的重复性和非终止性减少,并产生与最小风险训练相似的性能。
Abstract
neural autoregressive sequence models
are used to generate sequences in a variety of natural language processing (NLP) tasks, where they are evaluated according to sequence-level task losses. These models are typically trained with
→