Score-based Generative Models (SGMs) have achieved impressive performance in data generation across a wide range of applications and benefit from strong theoretical guarantees. Recently, methods inspired by statistical mechanics, in particular, Hamiltonian dynamics, have introduced Critically-damped Langevin Diffusions (CLDs), which define diffusion processes on extended spaces by coupling the data with auxiliary variables. These approaches, along with their associated score-matching and sampling procedures, have been shown to outperform standard diffusion-based samplers numerically. In this paper, we analyze a generalized dynamic that extends classical CLDs by introducing an additional hyperparameter controlling the noise applied to the data coordinate, thereby better exploiting the extended space. We further derive a novel upper bound on the sampling error of CLD-based generative models in the Wasserstein metric. This additional hyperparameter influences the smoothness of sample paths, and our discretization error analysis provides practical guidance for its tuning, leading to improved sampling performance.
翻译:基于分数的生成模型(SGMs)在广泛的数据生成应用中取得了卓越性能,并受益于坚实的理论保证。近期,受统计力学(特别是哈密顿动力学)启发的算法提出了临界阻尼朗之万扩散(CLDs),该方法通过在扩展空间中耦合数据与辅助变量来定义扩散过程。这些方法及其相关的分数匹配与采样流程,已在数值实验中展现出优于标准扩散采样器的性能。本文分析了一种广义动力学框架,通过引入控制数据坐标噪声的超参数扩展了经典CLD模型,从而更有效地利用扩展空间。我们进一步推导了基于CLD的生成模型在Wasserstein度量下采样误差的新上界。该超参数影响样本路径的光滑性,我们的离散化误差分析为其调优提供了实践指导,最终实现了采样性能的提升。