BriefGPT.xyz
Apr, 2019
通过知识蒸馏提高用于自然语言理解的多任务深度神经网络
Improving Multi-Task Deep Neural Networks via Knowledge Distillation for Natural Language Understanding
HTML
PDF
Xiaodong Liu, Pengcheng He, Weizhu Chen, Jianfeng Gao
TL;DR
本文使用知识蒸馏方法对多任务深度神经网络进行优化,提供了解决集成大型神经网络问题的方法,并在GLUE基准测试中取得了显著的成果。
Abstract
This paper explores the use of
knowledge distillation
to improve a Multi-Task
deep neural network
(MT-DNN) (Liu et al., 2019) for learning text representations across multiple
→