BriefGPT.xyz
Jul, 2022
使用私有联邦学习为资源受限设备训练大词汇神经语言模型
Training Large-Vocabulary Neural Language Models by Private Federated Learning for Resource-Constrained Devices
HTML
PDF
Mingbin Xu, Congzheng Song, Ye Tian, Neha Agrawal, Filip Granqvist...
TL;DR
使用联邦学习和差分隐私技术来保护隐私,同时采用PEU、LoRA和NCE等技术来降低大模型的噪声和内存需求,从而成功地在计算受限设备上训练大词汇量的语言模型。
Abstract
federated learning
(FL) is a technique to train models using data distributed across devices.
differential privacy
(DP) provides a formal privacy guarantee for sensitive data. Our goal is to train a large
→