Sponge examples are test-time inputs carefully-optimized to increase energy consumption and latency of neural networks when deployed on hardware accelerators. In this work, we demonstrate that sponge attacks can also be implanted at training time, when model training is outsourced to a third party, via an attack that we call sponge poisoning. This attack allows one to increase the energy consumption and latency of machine-learning models indiscriminately on each test-time input. We present a novel formalization for sponge poisoning, overcoming the limitations related to the optimization of test-time sponge examples, and show that this attack is possible even if the attacker only controls a few poisoning samples and model updates. Our extensive experimental analysis, involving two deep learning architectures and three datasets, shows that sponge poisoning can almost completely vanish the effect of such hardware accelerators. Finally, we analyze activations of the resulting sponge models, identifying the module components that are more sensitive to this vulnerability.
翻译:海绵的例子是指在硬件加速器上部署时仔细优化测试时间投入,以增加能量消耗和神经网络的静态。 在这项工作中,我们证明海绵袭击也可以在培训时间植入,在示范培训外包给第三方时,通过我们称之为海绵中毒的攻击,将示范培训外包给第三方。这次攻击允许一个人在每次试验时间投入上不加区别地增加机器学习模型的能量消耗和静态。我们提出了海绵中毒的新颖正规化,克服了与试验时间海绵范例优化有关的限制,并表明即使攻击者只控制少量中毒样品和模型更新,这种攻击也是可能的。我们涉及两个深层学习结构和三个数据集的广泛实验分析表明,海绵中毒几乎可以完全消除这种硬件加速器的效应。最后,我们分析了由此产生的海绵模型的启动情况,确定了对这一脆弱性更为敏感的模块组成部分。