Multi-level continuous encoding and decoding based on dilation convolution for super-resolution
Deep neural networks have shown better effects for super-resolution in recent years. However, it is difficult to extract multi-level features of low-resolution (LR) images to reconstruct more clear images. Most of the existing mainstream methods use encoding and decoding frameworks, which are still...
Gespeichert in:
Veröffentlicht in: | Multimedia tools and applications 2024-02, Vol.83 (7), p.20149-20167 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 20167 |
---|---|
container_issue | 7 |
container_start_page | 20149 |
container_title | Multimedia tools and applications |
container_volume | 83 |
creator | Zhang, Zhenghuan Ma, Yantu Liu, Wanjun Shi, Qiuhong |
description | Deep neural networks have shown better effects for super-resolution in recent years. However, it is difficult to extract multi-level features of low-resolution (LR) images to reconstruct more clear images. Most of the existing mainstream methods use encoding and decoding frameworks, which are still difficult to extract multi-level features from low resolution images, and this process is essential for the reconstruction of more clear images. To overcome these limitations, we present a multi-level continuous encoding and decoding based on dilation convolution for super-resolution (MEDSR). Specifically, we first construct a multi-level continuous encoding and decoding module, which can obtain more easy-to-extract features, complex-to-extract features, and difficult-to-extract features of LR images. Then we construct dilated attention modules based on different dilated rates to capture multi-level regional information of different respective fields and focus on each level information of multi-level regional information to extract multi-level deep features. These dilated attention modules are designed to incorporate varying levels of contextual information by dilating the receptive field of the attention module. This allows the module to attend to a larger area of the input while maintaining a constant memory footprint. MEDSR uses multi-level deep features of LR images to reconstruct better SR images, the values of PSNR and SSIM of our method on Set5 dataset reach 32.65 dB and 0.9005 respectively when the scale factor is ×4. Extensive experimental results demonstrate that our proposed MEDSR outperforms that of some state-of-the-art super-resolution methods. |
doi_str_mv | 10.1007/s11042-023-16415-5 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2927020622</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2927020622</sourcerecordid><originalsourceid>FETCH-LOGICAL-c270t-b9e275be0c3ac4c014bf5b106b13613957b4866a9c62ae0edebf80d12164566b3</originalsourceid><addsrcrecordid>eNp9UE1LxDAQDaLguvoHPBU8R2fSJt0eZfELFC96Dkk6XbrUZE3aBf-9cXdBT57mzfA-mMfYJcI1AtQ3CREqwUGUHFWFkssjNkNZl7yuBR7_wafsLKU1ACopqhnTL9Mw9nygLQ2FC37s_RSmVJB3oe39qjC-LVo6LNYkaovgi7YfzNhnkCXbMEw73IVYpGlDkUdKh-M5O-nMkOjiMOfs_f7ubfnIn18fnpa3z9yJGkZuGxK1tASuNK5ygJXtpEVQFkuFZSNrWy2UMo1TwhBQS7ZbQIsifyuVsuWcXe19NzF8TpRGvQ5T9DlSiyZHCFBCZJbYs1wMKUXq9Cb2HyZ-aQT9U6TeF6lzkXpXpJZZVO5FKZP9iuKv9T-qbyB-d0I</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2927020622</pqid></control><display><type>article</type><title>Multi-level continuous encoding and decoding based on dilation convolution for super-resolution</title><source>SpringerLink Journals</source><creator>Zhang, Zhenghuan ; Ma, Yantu ; Liu, Wanjun ; Shi, Qiuhong</creator><creatorcontrib>Zhang, Zhenghuan ; Ma, Yantu ; Liu, Wanjun ; Shi, Qiuhong</creatorcontrib><description>Deep neural networks have shown better effects for super-resolution in recent years. However, it is difficult to extract multi-level features of low-resolution (LR) images to reconstruct more clear images. Most of the existing mainstream methods use encoding and decoding frameworks, which are still difficult to extract multi-level features from low resolution images, and this process is essential for the reconstruction of more clear images. To overcome these limitations, we present a multi-level continuous encoding and decoding based on dilation convolution for super-resolution (MEDSR). Specifically, we first construct a multi-level continuous encoding and decoding module, which can obtain more easy-to-extract features, complex-to-extract features, and difficult-to-extract features of LR images. Then we construct dilated attention modules based on different dilated rates to capture multi-level regional information of different respective fields and focus on each level information of multi-level regional information to extract multi-level deep features. These dilated attention modules are designed to incorporate varying levels of contextual information by dilating the receptive field of the attention module. This allows the module to attend to a larger area of the input while maintaining a constant memory footprint. MEDSR uses multi-level deep features of LR images to reconstruct better SR images, the values of PSNR and SSIM of our method on Set5 dataset reach 32.65 dB and 0.9005 respectively when the scale factor is ×4. Extensive experimental results demonstrate that our proposed MEDSR outperforms that of some state-of-the-art super-resolution methods.</description><identifier>ISSN: 1573-7721</identifier><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-023-16415-5</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Algorithms ; Artificial neural networks ; Coding ; Computer Communication Networks ; Computer Science ; Convolution ; Data Structures and Information Theory ; Deep learning ; Image reconstruction ; Image resolution ; Methods ; Modules ; Multimedia ; Multimedia Information Systems ; Neural networks ; Special Purpose and Application-Based Systems</subject><ispartof>Multimedia tools and applications, 2024-02, Vol.83 (7), p.20149-20167</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2023. Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c270t-b9e275be0c3ac4c014bf5b106b13613957b4866a9c62ae0edebf80d12164566b3</cites><orcidid>0000-0002-2418-3986</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-023-16415-5$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-023-16415-5$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Zhang, Zhenghuan</creatorcontrib><creatorcontrib>Ma, Yantu</creatorcontrib><creatorcontrib>Liu, Wanjun</creatorcontrib><creatorcontrib>Shi, Qiuhong</creatorcontrib><title>Multi-level continuous encoding and decoding based on dilation convolution for super-resolution</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Deep neural networks have shown better effects for super-resolution in recent years. However, it is difficult to extract multi-level features of low-resolution (LR) images to reconstruct more clear images. Most of the existing mainstream methods use encoding and decoding frameworks, which are still difficult to extract multi-level features from low resolution images, and this process is essential for the reconstruction of more clear images. To overcome these limitations, we present a multi-level continuous encoding and decoding based on dilation convolution for super-resolution (MEDSR). Specifically, we first construct a multi-level continuous encoding and decoding module, which can obtain more easy-to-extract features, complex-to-extract features, and difficult-to-extract features of LR images. Then we construct dilated attention modules based on different dilated rates to capture multi-level regional information of different respective fields and focus on each level information of multi-level regional information to extract multi-level deep features. These dilated attention modules are designed to incorporate varying levels of contextual information by dilating the receptive field of the attention module. This allows the module to attend to a larger area of the input while maintaining a constant memory footprint. MEDSR uses multi-level deep features of LR images to reconstruct better SR images, the values of PSNR and SSIM of our method on Set5 dataset reach 32.65 dB and 0.9005 respectively when the scale factor is ×4. Extensive experimental results demonstrate that our proposed MEDSR outperforms that of some state-of-the-art super-resolution methods.</description><subject>Algorithms</subject><subject>Artificial neural networks</subject><subject>Coding</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Convolution</subject><subject>Data Structures and Information Theory</subject><subject>Deep learning</subject><subject>Image reconstruction</subject><subject>Image resolution</subject><subject>Methods</subject><subject>Modules</subject><subject>Multimedia</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Special Purpose and Application-Based Systems</subject><issn>1573-7721</issn><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>BENPR</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9UE1LxDAQDaLguvoHPBU8R2fSJt0eZfELFC96Dkk6XbrUZE3aBf-9cXdBT57mzfA-mMfYJcI1AtQ3CREqwUGUHFWFkssjNkNZl7yuBR7_wafsLKU1ACopqhnTL9Mw9nygLQ2FC37s_RSmVJB3oe39qjC-LVo6LNYkaovgi7YfzNhnkCXbMEw73IVYpGlDkUdKh-M5O-nMkOjiMOfs_f7ubfnIn18fnpa3z9yJGkZuGxK1tASuNK5ygJXtpEVQFkuFZSNrWy2UMo1TwhBQS7ZbQIsifyuVsuWcXe19NzF8TpRGvQ5T9DlSiyZHCFBCZJbYs1wMKUXq9Cb2HyZ-aQT9U6TeF6lzkXpXpJZZVO5FKZP9iuKv9T-qbyB-d0I</recordid><startdate>20240201</startdate><enddate>20240201</enddate><creator>Zhang, Zhenghuan</creator><creator>Ma, Yantu</creator><creator>Liu, Wanjun</creator><creator>Shi, Qiuhong</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-2418-3986</orcidid></search><sort><creationdate>20240201</creationdate><title>Multi-level continuous encoding and decoding based on dilation convolution for super-resolution</title><author>Zhang, Zhenghuan ; Ma, Yantu ; Liu, Wanjun ; Shi, Qiuhong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c270t-b9e275be0c3ac4c014bf5b106b13613957b4866a9c62ae0edebf80d12164566b3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Artificial neural networks</topic><topic>Coding</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Convolution</topic><topic>Data Structures and Information Theory</topic><topic>Deep learning</topic><topic>Image reconstruction</topic><topic>Image resolution</topic><topic>Methods</topic><topic>Modules</topic><topic>Multimedia</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Special Purpose and Application-Based Systems</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Zhenghuan</creatorcontrib><creatorcontrib>Ma, Yantu</creatorcontrib><creatorcontrib>Liu, Wanjun</creatorcontrib><creatorcontrib>Shi, Qiuhong</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Database (1962 - current)</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>ProQuest Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM global</collection><collection>Computing Database</collection><collection>ProQuest Research Library</collection><collection>Research Library (Corporate)</collection><collection>ProQuest advanced technologies & aerospace journals</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Zhenghuan</au><au>Ma, Yantu</au><au>Liu, Wanjun</au><au>Shi, Qiuhong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-level continuous encoding and decoding based on dilation convolution for super-resolution</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2024-02-01</date><risdate>2024</risdate><volume>83</volume><issue>7</issue><spage>20149</spage><epage>20167</epage><pages>20149-20167</pages><issn>1573-7721</issn><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Deep neural networks have shown better effects for super-resolution in recent years. However, it is difficult to extract multi-level features of low-resolution (LR) images to reconstruct more clear images. Most of the existing mainstream methods use encoding and decoding frameworks, which are still difficult to extract multi-level features from low resolution images, and this process is essential for the reconstruction of more clear images. To overcome these limitations, we present a multi-level continuous encoding and decoding based on dilation convolution for super-resolution (MEDSR). Specifically, we first construct a multi-level continuous encoding and decoding module, which can obtain more easy-to-extract features, complex-to-extract features, and difficult-to-extract features of LR images. Then we construct dilated attention modules based on different dilated rates to capture multi-level regional information of different respective fields and focus on each level information of multi-level regional information to extract multi-level deep features. These dilated attention modules are designed to incorporate varying levels of contextual information by dilating the receptive field of the attention module. This allows the module to attend to a larger area of the input while maintaining a constant memory footprint. MEDSR uses multi-level deep features of LR images to reconstruct better SR images, the values of PSNR and SSIM of our method on Set5 dataset reach 32.65 dB and 0.9005 respectively when the scale factor is ×4. Extensive experimental results demonstrate that our proposed MEDSR outperforms that of some state-of-the-art super-resolution methods.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-023-16415-5</doi><tpages>19</tpages><orcidid>https://orcid.org/0000-0002-2418-3986</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1573-7721 |
ispartof | Multimedia tools and applications, 2024-02, Vol.83 (7), p.20149-20167 |
issn | 1573-7721 1380-7501 1573-7721 |
language | eng |
recordid | cdi_proquest_journals_2927020622 |
source | SpringerLink Journals |
subjects | Algorithms Artificial neural networks Coding Computer Communication Networks Computer Science Convolution Data Structures and Information Theory Deep learning Image reconstruction Image resolution Methods Modules Multimedia Multimedia Information Systems Neural networks Special Purpose and Application-Based Systems |
title | Multi-level continuous encoding and decoding based on dilation convolution for super-resolution |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-14T17%3A16%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-level%20continuous%20encoding%20and%20decoding%20based%20on%20dilation%20convolution%20for%20super-resolution&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Zhang,%20Zhenghuan&rft.date=2024-02-01&rft.volume=83&rft.issue=7&rft.spage=20149&rft.epage=20167&rft.pages=20149-20167&rft.issn=1573-7721&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-023-16415-5&rft_dat=%3Cproquest_cross%3E2927020622%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2927020622&rft_id=info:pmid/&rfr_iscdi=true |