The continually advancing quality of deepfake technology exacerbates the threats of disinformation, fraud, and harassment by making maliciously-generated synthetic content increasingly difficult to distinguish from reality. We introduce a simple yet effective two-stage detection method that achieves an AUROC of over 99.8% on contemporary deepfakes. However, this high performance is short-lived. We show that models trained on this data suffer a recall drop of over 30% when evaluated on deepfakes created with generation techniques from just six months later, demonstrating significant decay as threats evolve. Our analysis reveals two key insights for robust detection. Firstly, continued performance requires the ongoing curation of large, diverse datasets. Second, predictive power comes primarily from static, frame-level artifacts, not temporal inconsistencies. The future of effective deepfake detection therefore depends on rapid data collection and the development of advanced frame-level feature detectors.
翻译:深度伪造技术的持续进步使得恶意生成的合成内容越来越难以与现实区分,从而加剧了虚假信息、欺诈和骚扰的威胁。我们提出了一种简单而有效的两阶段检测方法,在当代深度伪造内容上实现了超过99.8%的AUROC。然而,这种高性能是短暂的。我们发现,当使用仅六个月后更新的生成技术创建的深度伪造内容进行评估时,基于该数据训练的模型召回率下降超过30%,表明随着威胁演变出现了显著的性能衰减。我们的分析揭示了实现鲁棒检测的两个关键见解:首先,持续的性能表现需要不断构建大规模、多样化的数据集;其次,预测能力主要来自静态的帧级伪影,而非时间不一致性。因此,有效深度伪造检测的未来取决于快速的数据收集和先进的帧级特征检测器的开发。