Deep Neural Networks (DNN) are widely used in various fields due to their powerful performance, but recent studies have shown that deep learning models are vulnerable to adversarial attacks-by adding a slight perturbation to the input, the model will get wrong results. It is especially dangerous for some systems with high security requirements, so this paper proposes a new defense method by using the model super-fitting status. Model's adversarial robustness (i.e., the accuracry under adversarial attack) has been greatly improved in this status. This paper mathematically proves the effectiveness of super-fitting, and proposes a method to make the model reach this status quickly-minimaze unrelated categories scores (MUCS). Theoretically, super-fitting can resist any existing (even future) Based on CE white-box adversarial attack. In addition, this paper uses a variety of powerful attack algorithms to evaluate the adversarial robustness of super-fitting and other nearly 50 defense models from recent conferences. The experimental results show that super-fitting method in this paper can make the trained model obtain the highest adversarial performance robustness.
翻译:深神经网络(DNN)在各个领域被广泛使用,原因是其强大的性能,但最近的研究表明,深学习模式很容易受到对抗性攻击,因为输入会增加轻微的扰动,该模式将获得错误的结果。对于某些安全要求高的系统来说,它特别危险,因此本文件建议采用新的防御方法,使用模型超配状态。模型的对抗性强力(即对抗性攻击下的准确性)在这一状态下已大为改善。这份文件数学地证明了超装的有效性,并提出了使模型迅速达到这一状态的方法,即快速最小化不相关的类别分数(MCS)。理论上,超装可以抵制现有的(甚至未来)基于CE白箱对抗性攻击的任何(甚至未来)。此外,本文使用各种强大的攻击算法来评价超装和其他近50种防御模式的对抗性强力。实验结果表明,本文中的超配法方法可以使经过训练的模型获得最高对抗性强性能。