Model inversion (MI) attacks are aimed at reconstructing training data from model parameters. Such attacks have triggered increasing concerns about privacy, especially given a growing number of online model repositories. However, existing MI attacks against deep neural networks (DNNs) have large room for performance improvement. We present a novel inversion-specific GAN that can better distill knowledge useful for performing attacks on private models from public data. In particular, we train the discriminator to differentiate not only the real and fake samples but the soft-labels provided by the target model. Moreover, unlike previous work that directly searches for a single data point to represent a target class, we propose to model a private data distribution for each target class. Our experiments show that the combination of these techniques can significantly boost the success rate of the state-of-the-art MI attacks by 150%, and generalize better to a variety of datasets and models. Our code is available at https://github.com/SCccc21/Knowledge-Enriched-DMI.
翻译:模型反转(MI)攻击旨在从模型参数中重建培训数据。这种攻击引发了对隐私的越来越多的关注,特别是鉴于越来越多的在线模型库。然而,现有的对深神经网络的MI攻击有很大的改进性能的空间。我们提出了一个新的反转专用GAN,它可以更好地从公共数据中提取用于攻击私人模型的知识。特别是,我们训练歧视者不仅区分真实和假样品,而且区分目标模型提供的软标签。此外,与直接搜索一个数据点以代表一个目标类别的工作不同,我们提议为每个目标类别建立私人数据分发模型。我们的实验表明,这些技术的结合可以大大提高最先进的MI攻击的成功率150%,并更全面地推广到各种数据集和模型。我们的代码可在https://github.com/SCccc21/knowledgege-Enried-DMI上查阅。