BriefGPT.xyz
Apr, 2024
通过标签修订和数据选择改进知识蒸馏
Improve Knowledge Distillation via Label Revision and Data Selection
HTML
PDF
Weichao Lan, Yiu-ming Cheung, Qing Xu, Buhua Liu, Zhikai Hu...
TL;DR
该论文提出了一种解决知识蒸馏中错误监督的问题的方法,即通过标签修正纠正教师模型的错误预测,并引入数据选择技术以减少错误监督的影响,实验证明该方法的有效性,并表明该方法可以与其他蒸馏方法相结合,提高其性能。
Abstract
knowledge distillation
(KD) has become a widely used technique in the field of
model compression
, which aims to transfer knowledge from a large
t
→