论文标题
最佳任务分配到异构联合学习设备
Optimal Task Assignment to Heterogeneous Federated Learning Devices
论文作者
论文摘要
联合学习为培训机器学习模型提供了新的机会,同时尊重数据隐私。该技术基于异质设备,这些设备共同迭代训练模型,同时从未共享自己的数据。鉴于该培训的同步性质,联合学习系统的性能是由最慢的设备(也称为散散器)决定的。在本文中,我们通过控制每个设备用于训练的数据来最大程度地减少联合学习回合的持续时间的问题。我们将此问题提出为pan最小化问题,并具有相同的,独立和原子质的任务,这些任务必须分配给具有非降低成本功能的异质资源,同时尊重每个资源的任务的上限和上限。基于此公式,我们提出了一种名为Olar的多项式时间算法,并证明它提供了最佳的时间表。我们使用模拟在广泛的实验评估中评估了OLAR,该模拟包括与最新状态的其他算法进行比较以及对其新的扩展的比较。我们的结果表明,欧尔提供了较小的执行时间的最佳解决方案。他们还表明,每个资源的任务下限和上限的存在消除了次优启发式方法在算法执行时间方面可以提供的任何好处。
Federated Learning provides new opportunities for training machine learning models while respecting data privacy. This technique is based on heterogeneous devices that work together to iteratively train a model while never sharing their own data. Given the synchronous nature of this training, the performance of Federated Learning systems is dictated by the slowest devices, also known as stragglers. In this paper, we investigate the problem of minimizing the duration of Federated Learning rounds by controlling how much data each device uses for training. We formulate this problem as a makespan minimization problem with identical, independent, and atomic tasks that have to be assigned to heterogeneous resources with non-decreasing cost functions while respecting lower and upper limits of tasks per resource. Based on this formulation, we propose a polynomial-time algorithm named OLAR and prove that it provides optimal schedules. We evaluate OLAR in an extensive experimental evaluation using simulation that includes comparisons to other algorithms from the state of the art and new extensions to them. Our results indicate that OLAR provides optimal solutions with a small execution time. They also show that the presence of lower and upper limits of tasks per resource erase any benefits that suboptimal heuristics could provide in terms of algorithm execution time.