Federated Learning is a machine learning setting that reduces direct data exposure, improving the privacy guarantees of machine learning models. Yet, the exchange of model updates between the participants and the aggregator can still leak sensitive information. In this work, we present a new gradient-based membership inference attack for federated learning scenarios that exploits the temporal evolution of last-layer gradients across multiple federated rounds. Our method uses the shadow technique to learn round-wise gradient patterns of the training records, requiring no access to the private dataset, and is designed to consider both semi-honest and malicious adversaries (aggregators or data owners). Beyond membership inference, we also provide a natural extension of the proposed attack to discrete attribute inference by contrasting gradient responses under alternative attribute hypotheses. The proposed attacks are model-agnostic, and therefore applicable to any gradient-based model and can be applied to both classification and regression settings. We evaluate the attack on CIFAR-100 and Purchase100 datasets for membership inference and on Breast Cancer Wisconsin for attribute inference. Our findings reveal strong attack performance and comparable computational and memory overhead in membership inference when compared to another attack from the literature. The obtained results emphasize that multi-round federated learning can increase the vulnerability to inference attacks, that aggregators pose a more substantial threat than data owners, and that attack performance is strongly influenced by the nature of the training dataset, with richer, high-dimensional data leading to stronger leakage than simpler tabular data.
翻译:联邦学习是一种通过减少直接数据暴露来增强机器学习模型隐私保障的机器学习框架。然而,参与者与聚合器之间交换的模型更新仍可能泄露敏感信息。本文提出了一种新的基于梯度的联邦学习成员推理攻击方法,该方法利用多个联邦轮次中最后一层梯度的时序演化特征。我们的方法采用影子技术学习训练记录在每轮中的梯度模式,无需访问私有数据集,并设计用于同时考虑半诚实和恶意对手(聚合器或数据所有者)。除成员推理外,我们还通过对比不同属性假设下的梯度响应,将所提攻击自然扩展至离散属性推理任务。所提出的攻击方法具有模型无关性,因此适用于任何基于梯度的模型,并可应用于分类和回归场景。我们在CIFAR-100和Purchase100数据集上评估成员推理攻击效果,在威斯康星州乳腺癌数据集上评估属性推理性能。实验结果表明,与文献中的另一种攻击相比,本方法在成员推理任务中展现出强大的攻击性能,同时具有相当的计算与内存开销。研究结果强调:多轮联邦学习可能增加模型对推理攻击的脆弱性;聚合器比数据所有者构成更实质性的威胁;攻击性能受训练数据集特性显著影响,丰富的高维数据比简单的表格数据会导致更严重的信息泄露。