BriefGPT.xyz
Feb, 2020
理解和改善知识蒸馏
Understanding and Improving Knowledge Distillation
HTML
PDF
Jiaxi Tang, Rakesh Shivanna, Zhe Zhao, Dong Lin, Anima Singh...
TL;DR
本研究探讨了知识蒸馏的三个不同层次——宇宙,领域以及实例,发现这三个因素在知识蒸馏中起着重要作用,并在大量实证研究基础上,诊断了某些知识蒸馏应用失败的情况。
Abstract
knowledge distillation
is a model-agnostic technique to improve model quality while having a fixed capacity budget. It is a commonly used technique for
model compression
, where a higher capacity
→