Single image super-resolution with multi-scale information cross-fusion network

•A multi-scale cross module is developed to fuse multi-scale information and help information flow across the super-resolution network.•A residual-feature learning subnetwork with a cross-fusion structure is built to capture rich representations for image superresolution.•The cascaded structure is c...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Signal processing 2021-02, Vol.179, p.107831, Article 107831
Hauptverfasser: Hu, Yanting, Gao, Xinbo, Li, Jie, Huang, Yuanfei, Wang, Hanzi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:•A multi-scale cross module is developed to fuse multi-scale information and help information flow across the super-resolution network.•A residual-feature learning subnetwork with a cross-fusion structure is built to capture rich representations for image superresolution.•The cascaded structure is constructed and employed on several subnetworks to effectively recover details in a coarse-to-fine manner.•Extensive experiments are conducted to demonstrate the effectiveness and superiority of the proposed method. The deep convolutional neural networks have achieved significant improvements in terms of effectiveness and efficiency for single image super-resolution. However, as the depth of neural network grows, the information flow is weakened. On the other hand, most of the models adopt a single-stream structure with which integrating complementary information under different receptive fields is difficult. To improve information flow and capture sufficient knowledge for reconstructing high-frequency details, we propose a multi-scale information cross-fusion network (MSICF) in which a sequence of subnetworks is cascaded to infer high resolution features in a coarse-to-fine manner. In each cascaded subnetwork, we stack multiple multi-scale cross (MSC) modules to fuse multi-scale complementary information in an efficient way as well as to improve information flow across the layers. Meanwhile, by introducing residual-feature learning in each stage, the relative information between features at different levels is fully utilized to further boost reconstruction performance. We train the proposed network with cascaded-supervision and then assemble the intermediate predictions of the cascade to achieve high quality image. Extensive quantitative and qualitative evaluations on benchmark datasets illustrate the superiority of our proposed method over state-of-the-art super-resolution methods.
ISSN:0165-1684
1872-7557
DOI:10.1016/j.sigpro.2020.107831