Recent text-to-image diffusion models such as MidJourney and Stable Diffusion threaten to displace many in the professional artist community. In particular, models can learn to mimic the artistic style of specific artists after "fine-tuning" on samples of their art. In this paper, we describe the design, implementation and evaluation of Glaze, a tool that enables artists to apply "style cloaks" to their art before sharing online. These cloaks apply barely perceptible perturbations to images, and when used as training data, mislead generative models that try to mimic a specific artist. In coordination with the professional artist community, we deploy user studies to more than 1000 artists, assessing their views of AI art, as well as the efficacy of our tool, its usability and tolerability of perturbations, and robustness across different scenarios and against adaptive countermeasures. Both surveyed artists and empirical CLIP-based scores show that even at low perturbation levels (p=0.05), Glaze is highly successful at disrupting mimicry under normal conditions (>92%) and against adaptive countermeasures (>85%).
翻译:近期的文字到图像传播模型,如MidJourney 和 Snable Diflution 等,有可能取代专业艺术家界的许多人。 特别是, 模型可以学习模仿特定艺术家的艺术风格, 在其艺术样本上进行“ 微调” 。 在本文中, 我们描述了Glaze的设计、 实施和评估, 这个工具使艺术家能够在在线分享之前对艺术应用“ 风格斗篷 ” 。 这些斗篷几乎无法察觉地对图像进行扰动, 当用作培训数据时, 误导试图模仿某个特定艺术家的基因化模型。 我们与专业艺术家界协调, 向一千多名艺术家运用用户研究, 评估他们对AI艺术的看法, 以及我们工具的功效、 其可用性和可耐性、 在不同情况下的坚固性, 以及反对适应性反措施。 接受调查的艺术家和有经验的CLIP 得分都显示, 即使在低扰动级别( p=0.05 ), 格拉泽 也非常成功地在正常条件下破坏微缩缩缩缩( > 92% ) 和反对适应性反措施 ( > 85% )。