BriefGPT.xyz
Apr, 2024
句级别还是词级别?关于知识蒸馏的综合研究
Sentence-Level or Token-Level? A Comprehensive Study on Knowledge Distillation
HTML
PDF
Jingxuan Wei, Linzhuang Sun, Yichong Leng, Xu Tan, Bihui Yu...
TL;DR
通过改进的混合方法,将知识蒸馏技术应用于神经机器翻译,以在不同的情境中提高模型性能和压缩模型。
Abstract
knowledge distillation
, transferring knowledge from a teacher model to a student model, has emerged as a powerful technique in
neural machine translation
for compressing models or simplifying training targets.
→