Modern cameras' performance in low-light conditions remains suboptimal due to fundamental limitations in photon shot noise and sensor read noise. Generative image restoration methods have shown promising results compared to traditional approaches, but they suffer from hallucinatory content generation when the signal-to-noise ratio (SNR) is low. Leveraging the availability of personalized photo galleries of the users, we introduce Diffusion-based Personalized Generative Denoising (DiffPGD), a new approach that builds a customized diffusion model for individual users. Our key innovation lies in the development of an identity-consistent physical buffer that extracts the physical attributes of the person from the gallery. This ID-consistent physical buffer serves as a robust prior that can be seamlessly integrated into the diffusion model to restore degraded images without the need for fine-tuning. Over a wide range of low-light testing scenarios, we show that DiffPGD achieves superior image denoising and enhancement performance compared to existing diffusion-based denoising approaches. Our project page can be found at \href{https://genai-restore.github.io/DiffPGD/}{\textcolor{purple}{\textbf{https://genai-restore.github.io/DiffPGD/}}}.
翻译:由于光子散粒噪声和传感器读出噪声的根本性限制,现代相机在低光照条件下的性能仍不理想。与传统方法相比,生成式图像复原方法已展现出有前景的结果,但在信噪比(SNR)较低时,它们会因产生幻觉内容而受限。利用用户个性化照片库的可用性,我们提出了基于扩散的个性化生成式去噪(DiffPGD),这是一种为个体用户构建定制化扩散模型的新方法。我们的核心创新在于开发了一个身份一致物理缓冲区,该缓冲区从照片库中提取人物的物理属性。这一身份一致物理缓冲区作为一个鲁棒的先验,可无缝集成到扩散模型中,以恢复退化图像,而无需进行微调。在广泛的低光照测试场景中,我们证明DiffPGD相较于现有的基于扩散的去噪方法,实现了更优的图像去噪与增强性能。我们的项目页面位于:https://genai-restore.github.io/DiffPGD/。