Can Public Large Language Models Help Private Cross-device Federated Learning?
arxiv(2023)
摘要
We study (differentially) private federated learning (FL) of language models.
The language models in cross-device FL are relatively small, which can be
trained with meaningful formal user-level differential privacy (DP) guarantees
when massive parallelism in training is enabled by the participation of a
moderate size of users. Recently, public data has been used to improve
privacy-utility trade-offs for both large and small language models. In this
work, we provide a systematic study of using large-scale public data and LLMs
to help differentially private training of on-device FL models, and further
improve the privacy-utility tradeoff by techniques of distillation. Moreover,
we propose a novel distribution matching algorithm with theoretical grounding
to sample public data close to private data distribution, which significantly
improves the sample efficiency of (pre-)training on public data. The proposed
method is efficient and effective for training private models by taking
advantage of public data, especially for customized on-device architectures
that do not have ready-to-use pre-trained models.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要