论文标题
现实世界相机屏幕退化的联合生成学习和超分辨率
Joint Generative Learning and Super-Resolution For Real-World Camera-Screen Degradation
论文作者
论文摘要
在现实世界中的单图像超分辨率(SISR)任务中,低分辨率图像遭受了更复杂的降解,不仅被未知内核降采样。但是,通常使用合成的低分辨率生成(例如双孔插值(BI))研究了现有的SISR方法,这极大地限制了其性能。最近,一些研究人员从相机和智能手机的角度研究了现实世界中的SISR。但是,除了采购设备外,显示设备还涉及更复杂的降解。在本文中,我们专注于摄像机屏幕降解,并构建一个现实世界数据集(CAM-SCREENSR),其中HR图像是以前的Div2k数据集中的原始地面真相,相应的LR图像是屏幕上显示的HRS的相机捕获版本。我们进行了广泛的实验,以证明涉及更多实际降解对于改善SISR模型的概括是积极的。此外,我们提出了一个联合两阶段模型。首先,对降采样降解GAN(DD-GAN)进行了训练,可以对降解进行建模并产生更多的LR图像,该图像已验证,可有效地进行数据增强。然后,双重残留通道注意网络(Durcan)学会恢复SR图像。 L1损失和提议的拉普拉斯损失的加权组合用于锐化高频边缘。典型的合成和复杂的现实世界降解验证了所提出的方法比现有参数较少,更快的速度和更好的视觉结果的现实方法的广泛实验结果。此外,在实际捕获的照片中,我们的模型还提供最佳的视觉质量,具有更清晰的边缘,较少的伪像,尤其是适当的颜色增强功能,这尚未通过以前的方法完成。
In real-world single image super-resolution (SISR) task, the low-resolution image suffers more complicated degradations, not only downsampled by unknown kernels. However, existing SISR methods are generally studied with the synthetic low-resolution generation such as bicubic interpolation (BI), which greatly limits their performance. Recently, some researchers investigate real-world SISR from the perspective of the camera and smartphone. However, except the acquisition equipment, the display device also involves more complicated degradations. In this paper, we focus on the camera-screen degradation and build a real-world dataset (Cam-ScreenSR), where HR images are original ground truths from the previous DIV2K dataset and corresponding LR images are camera-captured versions of HRs displayed on the screen. We conduct extensive experiments to demonstrate that involving more real degradations is positive to improve the generalization of SISR models. Moreover, we propose a joint two-stage model. Firstly, the downsampling degradation GAN(DD-GAN) is trained to model the degradation and produces more various of LR images, which is validated to be efficient for data augmentation. Then the dual residual channel attention network (DuRCAN) learns to recover the SR image. The weighted combination of L1 loss and proposed Laplacian loss are applied to sharpen the high-frequency edges. Extensive experimental results in both typical synthetic and complicated real-world degradations validate the proposed method outperforms than existing SOTA models with less parameters, faster speed and better visual results. Moreover, in real captured photographs, our model also delivers best visual quality with sharper edge, less artifacts, especially appropriate color enhancement, which has not been accomplished by previous methods.