data-free knowledge distillation (DFKD) aims to obtain a lightweight student
model without original training data. Existing works generally synthesize data
from the pre-trained teacher model to replace the original training data for
student learning. To more effectively train the stude