Adversarial learning has been attracting more and more attention thanks to the fast development of machine learning and artificial intelligence. However, due to the complicated structure of most machine learning models, the mechanism of adversarial attacks is not well interpreted. How to measure the effect of attacks is still not quite clear. In this paper, we investigate the adversarial learning from the perturbation analysis point of view. We characterize the robustness of learning models through the calmness of the solution mapping. In the case of convex clustering models, we identify the conditions under which the clustering results remain the same under perturbations. When the noise level is large, it leads to an attack. Therefore, we propose two bilevel models for adversarial learning where the effect of adversarial learning is measured by some deviation function. Specifically, we systematically study the so-called $δ$-measure and show that under certain conditions, it can be used as a deviation function in adversarial learning for convex clustering models. Finally, we conduct numerical tests to verify the above theoretical results as well as the efficiency of the two proposed bilevel models.
翻译:随着机器学习和人工智能的快速发展,对抗学习已引起越来越多的关注。然而,由于大多数机器学习模型结构复杂,对抗攻击的机制尚未得到充分阐释。如何度量攻击的影响仍不甚明确。本文从扰动分析的角度研究对抗学习,通过解映射的镇定性来刻画学习模型的鲁棒性。针对凸聚类模型,我们确定了在扰动下聚类结果保持不变的条件。当噪声水平较大时,即构成攻击。为此,我们提出两种用于对抗学习的双层模型,其中对抗学习的效果通过偏差函数进行度量。具体而言,我们系统研究了所谓的$δ$-度量,并证明在一定条件下,该度量可作为凸聚类模型对抗学习中的偏差函数。最后,我们通过数值实验验证了上述理论结果以及所提两种双层模型的有效性。