BriefGPT.xyz
Sep, 2021
干细胞假说:利用Transformer编码器进行多任务学习面临的困境
The Stem Cell Hypothesis: Dilemma behind Multi-Task Learning with Transformer Encoders
HTML
PDF
Han He, Jinho D. Choi
TL;DR
通过多任务学习,结合transformer编码器,研究了多个NLP任务,发现多任务学习的注意力头之间相互干扰,提出干细胞假说解释在某些任务上具有天赋的注意力头不能被同时训练。同时,提出了新的无参考探针来验证该假说,通过标签分析展示了注意力头在五个任务之间如何被转换。
Abstract
multi-task learning
with
transformer encoders
(MTL) has emerged as a powerful technique to improve performance on closely-related tasks for both accuracy and efficiency while a question still remains whether or n
→