Unrestricted color attacks, which manipulate semantically meaningful color of an image, have shown their stealthiness and success in fooling both human eyes and deep neural networks. However, current works usually sacrifice the flexibility of the uncontrolled setting to ensure the naturalness of adversarial examples. As a result, the black-box attack performance of these methods is limited. To boost transferability of adversarial examples without damaging image quality, we propose a novel Natural Color Fool (NCF) which is guided by realistic color distributions sampled from a publicly available dataset and optimized by our neighborhood search and initialization reset. By conducting extensive experiments and visualizations, we convincingly demonstrate the effectiveness of our proposed method. Notably, on average, results show that our NCF can outperform state-of-the-art approaches by 15.0%$\sim$32.9% for fooling normally trained models and 10.0%$\sim$25.3% for evading defense methods. Our code is available at https://github.com/ylhz/Natural-Color-Fool.
翻译:不受限制的色彩攻击操纵了图像的内脏色彩,显示了其隐秘性和成功,欺骗了人类眼睛和深神经网络。然而,目前的工作通常牺牲不受控制的环境的灵活性,以确保对抗性实例的自然性。因此,这些方法的黑盒攻击性能受到限制。为了提高对抗性例子的可转让性,同时又不损害图像质量,我们提议了一个新型自然色彩愚昧(NCF),它以从公开可得的数据集取样的现实色彩分布为指导,并优化了我们社区搜索和初始化的重新设置。通过进行广泛的实验和可视化,我们令人信服地展示了我们拟议方法的有效性。值得注意的是,平均而言,结果显示我们的NCF能够以15.0%\sim32.9%的速度超越最先进的方法,以愚弄通常受过训练的模型和10.0%\sim 25.3%用于蒸发防御方法。我们的代码可在https://github.com/ylhz/Natural-Color-Fool查阅。