3D Gaussian Splatting (3DGS) has emerged as a state-of-the-art method for novel view synthesis. However, its performance heavily relies on dense, high-quality input imagery, an assumption that is often violated in real-world applications, where data is typically sparse and motion-blurred. These two issues create a vicious cycle: sparse views ignore the multi-view constraints necessary to resolve motion blur, while motion blur erases high-frequency details crucial for aligning the limited views. Thus, reconstruction often fails catastrophically, with fragmented views and a low-frequency bias. To break this cycle, we introduce CoherentGS, a novel framework for high-fidelity 3D reconstruction from sparse and blurry images. Our key insight is to address these compound degradations using a dual-prior strategy. Specifically, we combine two pre-trained generative models: a specialized deblurring network for restoring sharp details and providing photometric guidance, and a diffusion model that offers geometric priors to fill in unobserved regions of the scene. This dual-prior strategy is supported by several key techniques, including a consistency-guided camera exploration module that adaptively guides the generative process, and a depth regularization loss that ensures geometric plausibility. We evaluate CoherentGS through both quantitative and qualitative experiments on synthetic and real-world scenes, using as few as 3, 6, and 9 input views. Our results demonstrate that CoherentGS significantly outperforms existing methods, setting a new state-of-the-art for this challenging task. The code and video demos are available at https://potatobigroom.github.io/CoherentGS/.
翻译:3D高斯泼溅(3DGS)已成为新颖视角合成的先进方法。然而,其性能高度依赖于密集、高质量的输入图像,这一假设在实际应用中常被违背,因为数据通常稀疏且存在运动模糊。这两个问题形成了恶性循环:稀疏视角忽略了解决运动模糊所需的多视角约束,而运动模糊则抹除了对齐有限视角所必需的高频细节。因此,重建往往严重失败,表现为视角碎片化和低频偏差。为打破这一循环,我们提出了CoherentGS,这是一个从稀疏模糊图像进行高保真3D重建的新框架。我们的核心洞见在于采用双重先验策略应对这些复合退化。具体而言,我们结合了两个预训练的生成模型:一个专用于恢复清晰细节并提供光度指导的去模糊网络,以及一个为场景未观测区域提供几何先验的扩散模型。该双重先验策略得到多项关键技术的支持,包括自适应引导生成过程的一致性导向相机探索模块,以及确保几何合理性的深度正则化损失。我们通过在合成场景和真实场景上使用少至3、6、9个输入视角进行定量与定性实验来评估CoherentGS。结果表明,CoherentGS显著优于现有方法,为此挑战性任务设立了新的技术标杆。代码与视频演示可在https://potatobigroom.github.io/CoherentGS/获取。