We review the use of differential privacy (DP) for privacy protection in machine learning (ML). We show that, driven by the aim of preserving the accuracy of the learned models, DP-based ML implementations are so loose that they do not offer the ex ante privacy guarantees of DP. Instead, what they deliver is basically noise addition similar to the traditional (and often criticized) statistical disclosure control approach. Due to the lack of formal privacy guarantees, the actual level of privacy offered must be experimentally assessed ex post, which is done very seldom. In this respect, we present empirical results showing that standard anti-overfitting techniques in ML can achieve a better utility/privacy/efficiency trade-off than DP.
翻译:我们审查了在机器学习中使用不同隐私(DP)来保护隐私的情况。我们表明,由于旨在维护所学模式的准确性,基于DP的 ML实施过于松散,以致无法提供DP的事先隐私保障。相反,它们提供的基本上是噪音补充,类似于传统的(经常受到批评的)统计披露控制办法。由于缺乏正式的隐私保障,所提供的实际隐私水平必须在事后进行实验性评估,但很少这样做。在这方面,我们介绍了实证结果,表明ML的标准反改造技术能够比DP取得更好的效用/专利/效率交换。