Federated Learning (FL), despite demonstrating impressive capabilities in the training of multiple models in a decentralized manner, has been shown to produce a final model not necessarily well-suited to the needs of each client. While extensive work has been conducted on how to create tailored personalized models, called Personalized Federated Learning (PFL), less attention has been given to personalization via fine-tuning of foundation models with multi-task and multi-modal properties. Moreover, there exists a lack of understanding in the literature on how to fine-tune and personalize such models in a setting that is heterogeneous across clients not only in data, but also in tasks and modalities. To address this gap in the literature, we propose TAP (Two-Stage Adaptive Personalization), which (i) leverages mismatched model architectures between the clients and server to selectively conduct replacement operations when it benefits a client's local tasks and (ii) engages in post-FL knowledge distillation for capturing beneficial general knowledge without compromising personalization. We also introduce the first convergence analysis of the server model under its modality-task pair architecture, and demonstrate that as the number of modality-task pairs increases, its ability to cater to all tasks suffers. Through extensive experiments, we demonstrate the effectiveness of our proposed algorithm across a variety of datasets and tasks in comparison to a multitude of baselines. Implementation code is publicly available at https://github.com/lee3296/TAP.
翻译:尽管联邦学习(FL)在去中心化训练多个模型方面展现出卓越能力,但最终生成的模型未必能充分满足每个客户端的需求。虽然已有大量研究致力于创建定制化的个性化模型(即个性化联邦学习,PFL),但针对具有多任务与多模态特性的基础模型进行微调以实现个性化的研究尚不充分。此外,现有文献对如何在客户端间存在数据、任务及模态三重异质性的场景下微调并个性化此类模型仍缺乏深入理解。为填补这一研究空白,我们提出TAP(两阶段自适应个性化)方法,其具备以下特点:(i)利用客户端与服务器间的失配模型架构,在有利于客户端本地任务时选择性执行模型替换操作;(ii)通过联邦学习后阶段的知识蒸馏获取有益通用知识,同时保持个性化性能。我们首次对服务器模型在其模态-任务对架构下的收敛性进行了理论分析,证明随着模态-任务对数量增加,模型服务所有任务的能力将受到制约。通过大量实验,我们在多种数据集和任务场景中验证了所提算法相较于多种基线方法的优越性。实现代码已公开于https://github.com/lee3296/TAP。