论文标题
LotteryFL:非IID数据集上的彩票假设的个性化和沟通效率的联合学习
LotteryFL: Personalized and Communication-Efficient Federated Learning with Lottery Ticket Hypothesis on Non-IID Datasets
论文作者
论文摘要
联合学习是一种受欢迎的分布式机器学习范式,具有增强的隐私性。它的主要目标是学习一个全球模型,为尽可能多的参与者提供良好的表现。这项技术正在迅速发展,其中许多未解决的挑战,其中统计异质性(即非IID)和沟通效率是两种关键的挑战,这些挑战阻碍了联合学习的发展。在这项工作中,我们提出了Lotteryfl,这是一个通过利用彩票假设来实现个性化和沟通效率的联合学习框架。在Lotteryfl中,每个客户通过应用彩票假设来学习彩票网络(即基本模型的子网),并且仅在服务器和客户端之间传达这些彩票网络。每个客户都没有在经典联邦学习中学习共享的全球模型,而是通过Lotteryfl学习个性化模型。由于彩票网络的紧凑型,可以大大降低通信成本。为了支持我们框架的培训和评估,我们通过使用功能分布偏斜,标签分布偏斜和数量偏斜来构建基于MNIST,CIFAR-10和EMNIST的非IID数据集。这些非IID数据集的实验表明,从个性化和沟通成本方面,Lotteryfl明显优于现有的解决方案。
Federated learning is a popular distributed machine learning paradigm with enhanced privacy. Its primary goal is learning a global model that offers good performance for the participants as many as possible. The technology is rapidly advancing with many unsolved challenges, among which statistical heterogeneity (i.e., non-IID) and communication efficiency are two critical ones that hinder the development of federated learning. In this work, we propose LotteryFL -- a personalized and communication-efficient federated learning framework via exploiting the Lottery Ticket hypothesis. In LotteryFL, each client learns a lottery ticket network (i.e., a subnetwork of the base model) by applying the Lottery Ticket hypothesis, and only these lottery networks will be communicated between the server and clients. Rather than learning a shared global model in classic federated learning, each client learns a personalized model via LotteryFL; the communication cost can be significantly reduced due to the compact size of lottery networks. To support the training and evaluation of our framework, we construct non-IID datasets based on MNIST, CIFAR-10 and EMNIST by taking feature distribution skew, label distribution skew and quantity skew into consideration. Experiments on these non-IID datasets demonstrate that LotteryFL significantly outperforms existing solutions in terms of personalization and communication cost.