论文标题

仅从未标记的数据中与班级共享客户端进行联合学习

Federated Learning from Only Unlabeled Data with Class-Conditional-Sharing Clients

论文作者

Lu, Nan, Wang, Zhao, Li, Xiaoxiao, Niu, Gang, Dou, Qi, Sugiyama, Masashi

论文摘要

监督联邦学习(FL)使多个客户可以共享经过训练的模型而无需共享标签的数据。但是,潜在客户甚至可能不愿意标记自己的数据,这可能会限制FL在实践中的适用性。在本文中,我们展示了无监督的FL的可能性,其模型仍然是预测类标签的分类器,如果类别的概率发生了变化,而同类条件分布在客户端拥有的未标记数据中共享。我们提出了无监督学习的联合会(Fedul),其中未标记的数据被转换为每个客户端的替代标记数据,经过修改的模型经过监督的FL培训,并且从修改的模型中恢复了通缉模型。 Fedul是无监督的FL的非常通用的解决方案:它与许多监督的FL方法兼容,并且可以从理论上保证所需模型的恢复,就好像数据已被标记一样。基准和现实世界数据集的实验证明了Fedul的有效性。代码可在https://github.com/lunanbit/fedul上找到。

Supervised federated learning (FL) enables multiple clients to share the trained model without sharing their labeled data. However, potential clients might even be reluctant to label their own data, which could limit the applicability of FL in practice. In this paper, we show the possibility of unsupervised FL whose model is still a classifier for predicting class labels, if the class-prior probabilities are shifted while the class-conditional distributions are shared among the unlabeled data owned by the clients. We propose federation of unsupervised learning (FedUL), where the unlabeled data are transformed into surrogate labeled data for each of the clients, a modified model is trained by supervised FL, and the wanted model is recovered from the modified model. FedUL is a very general solution to unsupervised FL: it is compatible with many supervised FL methods, and the recovery of the wanted model can be theoretically guaranteed as if the data have been labeled. Experiments on benchmark and real-world datasets demonstrate the effectiveness of FedUL. Code is available at https://github.com/lunanbit/FedUL.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源