BriefGPT.xyz
Feb, 2023
利用预训练语言模型稳定上下文学习以实现少样本对话状态跟踪
Stabilized In-Context Learning with Pre-trained Language Models for Few Shot Dialogue State Tracking
HTML
PDF
Derek Chen, Kun Qian, Zhou Yu
TL;DR
通过元学习在对话领域稳定模型,设计新的训练方法改进检索机制寻找理想的示例,在句子长度有限的情况下,使用显著性模型限制对话文本长度,从而在MultiWOZ上实现了具有竞争力的少样本对话状态跟踪结果。
Abstract
Prompt-based methods with large
pre-trained language models
(PLMs) have shown impressive unaided performance across many NLP tasks. These models improve even further with the addition of a few labeled
in-context exempla
→