Data removal from machine learning models has been paid more attentions due to the demands of the "right to be forgotten" and countering data poisoning attacks. In this paper, we frame the problem of federated unlearning, a post-process operation of the federated learning models to remove the influence of the specified training sample(s). We present FedEraser, the first federated unlearning methodology that can eliminate the influences of a federated client's data on the global model while significantly reducing the time consumption used for constructing the unlearned model. The core idea of FedEraser is to trade the central server's storage for unlearned model's construction time. In particular, FedEraser reconstructs the unlearned model by leveraging the historical parameter updates of federated clients that have been retained at the central server during the training process of FL. A novel calibration method is further developed to calibrate the retained client updates, which can provide a significant speed-up to the reconstruction of the unlearned model. Experiments on four realistic datasets demonstrate the effectiveness of FedEraser, with an expected speed-up of $4\times$ compared with retraining from the scratch.
翻译:由于“被遗忘的权利”和打击数据中毒袭击的要求,机器学习模型的数据去除工作受到更多关注。在本文中,我们界定了联合不学习的问题,即联合学习模型的后处理操作,以消除特定培训样本的影响。我们介绍了FedEraser,这是第一个能够消除联合客户数据对全球模式的影响的联合会不学习方法,同时大大缩短构建未学习模型所用时间的消耗。FedEraser的核心理念是将中央服务器存储器用于交换未学习模型的建设时间。特别是,FedEraser利用FL培训过程中保留在中央服务器上的联邦用户的历史参数更新来重建未学习模型。正在进一步开发新校准方法,以校准保留客户数据更新,这可以为重建未学习模型提供重要的速度。在四个现实的数据集上进行的实验显示了FedEraser的实效性,预计从重新培训中将获得4美元的速度提高到4美元。