Image super-resolution using very deep residual channel attention networks

Y Zhang, K Li, K Li, L Wang… - Proceedings of the …, 2018 - openaccess.thecvf.com
Proceedings of the European conference on computer vision (ECCV), 2018openaccess.thecvf.com
Convolutional neural network (CNN) depth is of crucial importance for image super-
resolution (SR). However, we observe that deeper networks for image SR are more difficult
to train. The low-resolution (LR) inputs and features contain abundant low-frequency
information, which is treated equally across channels, hence hindering the representational
ability of CNNs. To solve these problems, we propose the very deep residual channel
attention networks (RCAN). Specifically, we propose residual in residual (RIR) structure to …
Abstract
Convolutional neural network (CNN) depth is of crucial importance for image super-resolution (SR). However, we observe that deeper networks for image SR are more difficult to train. The low-resolution (LR) inputs and features contain abundant low-frequency information, which is treated equally across channels, hence hindering the representational ability of CNNs. To solve these problems, we propose the very deep residual channel attention networks (RCAN). Specifically, we propose residual in residual (RIR) structure to form very deep network, which consists of several residual groups with long skip connections. Each residual group contains some residual blocks with short skip connections. Meanwhile, RIR allows abundant low-frequency information to be bypassed through multiple skip connections, making the main network focus on learning high-frequency information. Furthermore, we propose channel attention mechanism to adaptively rescale channel-wise features by considering interdependencies among channels. Extensive experiments show that our RCAN achieves better accuracy and visual improvements against state-of-the-art methods.
openaccess.thecvf.com
以上显示的是最相近的搜索结果。 查看全部搜索结果