论文标题

SCSNET:同时学习图像着色和超分辨率的有效范式

SCSNet: An Efficient Paradigm for Learning Simultaneously Image Colorization and Super-Resolution

论文作者

Zhang, Jiangning, Xu, Chao, Li, Jian, Han, Yue, Wang, Yabiao, Tai, Ying, Liu, Yong

论文摘要

在恢复低分辨率灰度图像的实际应用中,我们通常需要为目标设备运行三个单独的图像着色,超分辨率和DOWS采样操作的过程。但是,对于独立过程,该管道是冗余且效率低下的,并且可以共享某些内部功能。因此,我们提出了一个有效的范式,以执行{s}图像{c}浮力化和{s} uper-resolution(SCS),并提出一个端到端的SCSNET来实现此目标。所提出的方法由两个部分组成:用于学习颜色信息的着色分支,用于采用所提出的插件\ emph {金字塔阀交叉注意}(PVCATTN)模块,以在源图像和参考图像之间汇总特征图;和用于集成颜色和纹理信息的超分辨率分支以预测目标图像,该图像使用设计的\ emph {连续像素映射}(CPM)模块在连续放大倍率下预测高分辨率图像。此外,我们的SCSNET支持自动和参照模式,这些模式更灵活,用于实际应用。 Abundant experiments demonstrate the superiority of our method for generating authentic images over state-of-the-art methods, e.g., averagely decreasing FID by 1.8$\downarrow$ and 5.1 $\downarrow$ compared with current best scores for automatic and referential modes, respectively, while owning fewer parameters (more than $\times$2$\downarrow$) and faster running speed (more than $ \ times $ 3 $ \ uparrow $)。

In the practical application of restoring low-resolution gray-scale images, we generally need to run three separate processes of image colorization, super-resolution, and dows-sampling operation for the target device. However, this pipeline is redundant and inefficient for the independent processes, and some inner features could have been shared. Therefore, we present an efficient paradigm to perform {S}imultaneously Image {C}olorization and {S}uper-resolution (SCS) and propose an end-to-end SCSNet to achieve this goal. The proposed method consists of two parts: colorization branch for learning color information that employs the proposed plug-and-play \emph{Pyramid Valve Cross Attention} (PVCAttn) module to aggregate feature maps between source and reference images; and super-resolution branch for integrating color and texture information to predict target images, which uses the designed \emph{Continuous Pixel Mapping} (CPM) module to predict high-resolution images at continuous magnification. Furthermore, our SCSNet supports both automatic and referential modes that is more flexible for practical application. Abundant experiments demonstrate the superiority of our method for generating authentic images over state-of-the-art methods, e.g., averagely decreasing FID by 1.8$\downarrow$ and 5.1 $\downarrow$ compared with current best scores for automatic and referential modes, respectively, while owning fewer parameters (more than $\times$2$\downarrow$) and faster running speed (more than $\times$3$\uparrow$).

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源