BriefGPT.xyz
Mar, 2019
从BERT中提炼出特定任务的知识并转化为简单的神经网络
Distilling Task-Specific Knowledge from BERT into Simple Neural Networks
HTML
PDF
Raphael Tang, Yao Lu, Linqing Liu, Lili Mou, Olga Vechtomova...
TL;DR
本文主要研究神经网络中的自然语言处理,通过将BERT知识蒸馏成单层BiLSTM及其同侧对应的句对任务,证明了浅层神经网络仍然可以在不使用架构变化、外部训练数据或其他输入特征的情况下,与ELMo相比获得可媲美的结果。
Abstract
In the
natural language processing
literature,
neural networks
are becoming increasingly deeper and complex. The recent poster child of this trend is the deep
→