SCALA: Split Federated Learning with Concatenated Activations and Logit Adjustments
arxiv(2024)
摘要
Split Federated Learning (SFL) is a distributed machine learning framework
which strategically divides the learning process between a server and clients
and collaboratively trains a shared model by aggregating local models updated
based on data from distributed clients. However, data heterogeneity and partial
client participation result in label distribution skew, which severely degrades
the learning performance. To address this issue, we propose SFL with
Concatenated Activations and Logit Adjustments (SCALA). Specifically, the
activations from the client-side models are concatenated as the input of the
server-side model so as to centrally adjust label distribution across different
clients, and logit adjustments of loss functions on both server-side and
client-side models are performed to deal with the label distribution variation
across different subsets of participating clients. Theoretical analysis and
experimental results verify the superiority of the proposed SCALA on public
datasets.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要