BriefGPT.xyz
May, 2024
课程数据蒸馏
Curriculum Dataset Distillation
HTML
PDF
Zhiheng Ma, Anjia Cao, Funing Yang, Xing Wei
TL;DR
本论文提出了一个基于课程设置的数据集蒸馏框架,通过将从简单到复杂的课程进行逐步蒸馏,合理处理合成图像的生成和评估,进一步利用对抗优化改进图像的代表性,提高其在不同神经网络体系结构下的泛化能力和抗噪性,实现了大规模数据集蒸馏的新突破。
Abstract
Most
dataset distillation
methods struggle to accommodate large-scale datasets due to their substantial computational and memory requirements. In this paper, we present a
curriculum-based
→