BriefGPT.xyz
Oct, 2023
基于知识蒸馏的异构联邦学习
Heterogeneous Federated Learning Using Knowledge Codistillation
HTML
PDF
Jared Lichtarge, Ehsan Amid, Shankar Kumar, Tien-Ju Yang, Rohan Anil...
TL;DR
通过使用双向知识蒸馏方法,在具有不同性能的一部分客户端上训练较大的模型和整体池上训练较小的模型,实现不同领域之间的模型域转移,从而提高联邦平均算法的性能。该方法在图像分类和语言建模任务中表现出改进的效果,即使只有领域外或领域内有限的蒸馏数据可用。
Abstract
federated averaging
, and many federated learning algorithm variants which build upon it, have a limitation: all clients must share the same
model architecture
. This results in unused modeling capacity on many cli
→