In federated learning, models are learned from users' data that are held private in their edge devices, by aggregating them in the service provider's "cloud" to obtain a global model. Such global model is of great commercial value in, e.g., improving the customers' experience. In this paper we focus on two possible areas of improvement of the state of the art. First, we take the difference between user habits into account and propose a quadratic penalty-based formulation, for efficient learning of the global model that allows to personalize local models. Second, we address the latency issue associated with the heterogeneous training time on edge devices, by exploiting a hierarchical structure modeling communication not only between the cloud and edge devices, but also within the cloud. Specifically, we devise a tailored block coordinate descent-based computation scheme, accompanied with communication protocols for both the synchronous and asynchronous cloud settings. We characterize the theoretical convergence rate of the algorithm, and provide a variant that performs empirically better. We also prove that the asynchronous protocol, inspired by multi-agent consensus technique, has the potential for large gains in latency compared to a synchronous setting when the edge-device updates are intermittent. Finally, experimental results are provided that corroborate not only the theory, but also show that the system leads to faster convergence for personalized models on the edge devices, compared to the state of the art.
翻译:在联结式学习中,从用户的数据中学习模型,这些模型在边缘设备中是私密的,方法是将它们集中到服务供应商的“球球”中,以获得全球模型。这种全球模型具有巨大的商业价值,例如,改善客户的经验。在本文中,我们侧重于改进艺术状态的两个可能领域。首先,我们考虑到用户习惯之间的差异,并提议一种基于二次惩罚的公式,以便有效地学习能够使当地模型个人化的全球模型。第二,我们处理与边缘设备不同培训时间相关的悬浮问题,利用等级结构模型,不仅在云和边缘设备之间,而且在云中进行。具体地说,我们设计了一个有针对性的区块,协调基于世系的计算办法,同时结合同步和不同步的云层环境的通信协议。我们描述算法的理论趋同率,并提供一个实验性更好的变式。我们还证明,在多剂共识技术的启发下,在边缘设备上,通过利用等级结构建模的通信,不仅在云和边缘设备之间,而且在云中进行大型的趋同率,还有可能使系统取得巨大的趋同。