BriefGPT.xyz
Jun, 2023
论注意力在提示调整中的作用
On the Role of Attention in Prompt-tuning
HTML
PDF
Samet Oymak, Ankit Singh Rawat, Mahdi Soltanolkotabi, Christos Thrampoulidis
TL;DR
本文针对单层的注意力机制和上下文混合模型, 探究了Prompt-tuning的策略, 并分析了Prompt在梯度下降过程中学到的规律和其在上下文中的表现, 最终提出了有关Prompt-tuning的理论解释和实验验证。
Abstract
prompt-tuning
is an emerging strategy to adapt large
language models
(LLM) to downstream tasks by learning a (soft-)prompt parameter from data. Despite its success in LLMs, there is limited theoretical understand
→