Image correction and rectangling are valuable tasks in practical photography systems such as smartphones. Recent remarkable advancements in deep learning have undeniably brought about substantial performance improvements in these fields. Nevertheless, existing methods mainly rely on task-specific architectures. This significantly restricts their generalization ability and effective application across a wide range of different tasks. In this paper, we introduce the Unified Rectification Framework (UniRect), a comprehensive approach that addresses these practical tasks from a consistent distortion rectification perspective. Our approach incorporates various task-specific inverse problems into a general distortion model by simulating different types of lenses. To handle diverse distortions, UniRect adopts one task-agnostic rectification framework with a dual-component structure: a {Deformation Module}, which utilizes a novel Residual Progressive Thin-Plate Spline (RP-TPS) model to address complex geometric deformations, and a subsequent Restoration Module, which employs Residual Mamba Blocks (RMBs) to counteract the degradation caused by the deformation process and enhance the fidelity of the output image. Moreover, a Sparse Mixture-of-Experts (SMoEs) structure is designed to circumvent heavy task competition in multi-task learning due to varying distortions. Extensive experiments demonstrate that our models have achieved state-of-the-art performance compared with other up-to-date methods.
翻译:图像校正与矩形化是智能手机等实用摄影系统中的重要任务。深度学习近年来的显著进展无疑为这些领域带来了实质性的性能提升。然而,现有方法主要依赖于任务特定的架构,这极大地限制了其泛化能力及在广泛不同任务中的有效应用。本文提出统一校正框架(UniRect),这是一种从一致的畸变校正视角处理这些实际任务的综合方法。我们的方法通过模拟不同类型的镜头,将各种任务特定的逆问题整合到通用畸变模型中。为处理多样化的畸变,UniRect采用了一个具有双组件结构的任务无关校正框架:{变形模块}利用新颖的残差渐进薄板样条(RP-TPS)模型处理复杂几何变形;随后的恢复模块采用残差Mamba块(RMBs)来抵消变形过程导致的退化并提升输出图像的保真度。此外,我们设计了稀疏专家混合(SMoEs)结构,以规避因畸变差异导致的多任务学习中严重的任务竞争问题。大量实验表明,与当前最新方法相比,我们的模型取得了最先进的性能。