SCALR: Communication-Efficient Secure Multi-Party Logistic Regression.

IEEE Trans. Commun.(2024)

引用 0|浏览0
暂无评分
摘要
Privacy-preserving coded computing is a popular framework for multiple data-owners to jointly train machine learning models, with strong end-to-end information-theoretic privacy guarantees for the local data. A major challenge against the scalability of current approaches is their communication overhead, which is quadratic in the number of users. Towards addressing this challenge, we present SCALR, a communication-efficient collaborative learning framework for training logistic regression models. To do so, we introduce a novel coded computing mechanism, by decoupling the communication-intensive encoding operations from real-time training, and offloading the former to a data-independent offline phase, where the communicated variables are independent from training data. As such, the offline phase can be executed proactively during periods of low network activity. Communication complexity of the data-dependent (online) training operations is only linear in the number of users, greatly reducing the quadratic state-of-the-art. Our theoretical analysis presents the information-theoretic privacy guarantees, and shows that SCALR achieves the same performance guarantees as the state-of-the-art, in terms of adversary resilience, robustness to user dropouts, and model convergence. Through extensive experiments, we demonstrate up to $80\times $ reduction in online communication overhead, and $6\times $ speed-up in the wall-clock training time compared to the state-of-the-art.
更多
查看译文
关键词
Privacy-preserving distributed learning,information-theory,decentralized training
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要