Multiscale Attentive Image De-Raining Networks via Neural Architecture Search

Multi-scale architectures and attention modules have shown effectiveness in many deep learning-based image de-raining methods. However, manually designing and integrating these two components into a neural network requires a bulk of labor and extensive expertise. In this article, a high-performance...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on circuits and systems for video technology 2023-02, Vol.33 (2), p.618-633
Hauptverfasser: Cai, Lei, Fu, Yuli, Huo, Wanliang, Xiang, Youjun, Zhu, Tao, Zhang, Ying, Zeng, Huanqiang, Zeng, Delu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 633
container_issue 2
container_start_page 618
container_title IEEE transactions on circuits and systems for video technology
container_volume 33
creator Cai, Lei
Fu, Yuli
Huo, Wanliang
Xiang, Youjun
Zhu, Tao
Zhang, Ying
Zeng, Huanqiang
Zeng, Delu
description Multi-scale architectures and attention modules have shown effectiveness in many deep learning-based image de-raining methods. However, manually designing and integrating these two components into a neural network requires a bulk of labor and extensive expertise. In this article, a high-performance multi-scale attentive neural architecture search (MANAS) framework is technically developed for image de-raining. The proposed method formulates a new multi-scale attention search space with multiple flexible modules that are favorite to the image de-raining task. Under the search space, multi-scale attentive cells are built, which are further used to construct a powerful image de-raining network. The internal multi-scale attentive architecture of the de-raining network is searched automatically through a gradient-based search algorithm, which avoids the daunting procedure of the manual design to some extent. Moreover, in order to obtain a robust image de-raining model, a practical and effective multi-to- one training strategy is also presented to allow the de-raining network to get sufficient background information from multiple rainy images with the same background scene, and meanwhile, multiple loss functions including external loss, internal loss, architecture regularization loss, and model complexity loss are jointly optimized to achieve robust de-raining performance and controllable model complexity. Extensive experimental results on both synthetic and realistic rainy images, as well as the down-stream vision applications (i.e., objection detection and segmentation) consistently demonstrate the superiority of our proposed method. The code is publicly available at https://github.com/lcai-gz/MANAS .
doi_str_mv 10.1109/TCSVT.2022.3207516
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_journals_2773447993</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9894375</ieee_id><sourcerecordid>2773447993</sourcerecordid><originalsourceid>FETCH-LOGICAL-c295t-52789d78e2ed59dabd6660c5efcc24d00af4021af2c36d73e9a353104102259e3</originalsourceid><addsrcrecordid>eNo9kEtPwzAQhC0EEqXwB-ASiXOKn3F8jMqrUgsSLVwt42yKS5oU2yni35PSitPuSDO7mg-hS4JHhGB1sxjP3xYjiikdMYqlINkRGhAh8pRSLI77HQuS5pSIU3QWwgpjwnMuB2g26-rogjU1JEWM0ES3hWSyNktIbiF9Ma5xzTJ5gvjd-s-QbJ3pRedNnRTefrgINnYekjmYXp6jk8rUAS4Oc4he7-8W48d0-vwwGRfT1FIlYiqozFUpc6BQClWa9zLLMmwFVNZSXmJsKo4pMRW1LCslA2WYYARz0hcUCtgQXe_vbnz71UGIetV2vulfaiol41wqxXoX3busb0PwUOmNd2vjfzTBeodN_2HTO2z6gK0PXe1DDgD-AypXnEnBfgFHd2ii</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2773447993</pqid></control><display><type>article</type><title>Multiscale Attentive Image De-Raining Networks via Neural Architecture Search</title><source>IEEE Electronic Library (IEL)</source><creator>Cai, Lei ; Fu, Yuli ; Huo, Wanliang ; Xiang, Youjun ; Zhu, Tao ; Zhang, Ying ; Zeng, Huanqiang ; Zeng, Delu</creator><creatorcontrib>Cai, Lei ; Fu, Yuli ; Huo, Wanliang ; Xiang, Youjun ; Zhu, Tao ; Zhang, Ying ; Zeng, Huanqiang ; Zeng, Delu</creatorcontrib><description>Multi-scale architectures and attention modules have shown effectiveness in many deep learning-based image de-raining methods. However, manually designing and integrating these two components into a neural network requires a bulk of labor and extensive expertise. In this article, a high-performance multi-scale attentive neural architecture search (MANAS) framework is technically developed for image de-raining. The proposed method formulates a new multi-scale attention search space with multiple flexible modules that are favorite to the image de-raining task. Under the search space, multi-scale attentive cells are built, which are further used to construct a powerful image de-raining network. The internal multi-scale attentive architecture of the de-raining network is searched automatically through a gradient-based search algorithm, which avoids the daunting procedure of the manual design to some extent. Moreover, in order to obtain a robust image de-raining model, a practical and effective multi-to- one training strategy is also presented to allow the de-raining network to get sufficient background information from multiple rainy images with the same background scene, and meanwhile, multiple loss functions including external loss, internal loss, architecture regularization loss, and model complexity loss are jointly optimized to achieve robust de-raining performance and controllable model complexity. Extensive experimental results on both synthetic and realistic rainy images, as well as the down-stream vision applications (i.e., objection detection and segmentation) consistently demonstrate the superiority of our proposed method. The code is publicly available at https://github.com/lcai-gz/MANAS .</description><identifier>ISSN: 1051-8215</identifier><identifier>EISSN: 1558-2205</identifier><identifier>DOI: 10.1109/TCSVT.2022.3207516</identifier><identifier>CODEN: ITCTEM</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Complexity ; Complexity theory ; Computer architecture ; Controllability ; Image de-raining ; Image segmentation ; Machine learning ; Modules ; multi-scale attentive neural architecture search ; multi-to-one training strategy ; Neural networks ; Rain ; Regularization ; Robust control ; Search algorithms ; Shape ; Task analysis ; Training</subject><ispartof>IEEE transactions on circuits and systems for video technology, 2023-02, Vol.33 (2), p.618-633</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c295t-52789d78e2ed59dabd6660c5efcc24d00af4021af2c36d73e9a353104102259e3</citedby><cites>FETCH-LOGICAL-c295t-52789d78e2ed59dabd6660c5efcc24d00af4021af2c36d73e9a353104102259e3</cites><orcidid>0000-0002-2109-948X ; 0000-0002-6548-446X ; 0000-0001-7322-1873 ; 0000-0002-2802-7745 ; 0000-0003-4763-3529 ; 0000-0002-3244-6459</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9894375$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9894375$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Cai, Lei</creatorcontrib><creatorcontrib>Fu, Yuli</creatorcontrib><creatorcontrib>Huo, Wanliang</creatorcontrib><creatorcontrib>Xiang, Youjun</creatorcontrib><creatorcontrib>Zhu, Tao</creatorcontrib><creatorcontrib>Zhang, Ying</creatorcontrib><creatorcontrib>Zeng, Huanqiang</creatorcontrib><creatorcontrib>Zeng, Delu</creatorcontrib><title>Multiscale Attentive Image De-Raining Networks via Neural Architecture Search</title><title>IEEE transactions on circuits and systems for video technology</title><addtitle>TCSVT</addtitle><description>Multi-scale architectures and attention modules have shown effectiveness in many deep learning-based image de-raining methods. However, manually designing and integrating these two components into a neural network requires a bulk of labor and extensive expertise. In this article, a high-performance multi-scale attentive neural architecture search (MANAS) framework is technically developed for image de-raining. The proposed method formulates a new multi-scale attention search space with multiple flexible modules that are favorite to the image de-raining task. Under the search space, multi-scale attentive cells are built, which are further used to construct a powerful image de-raining network. The internal multi-scale attentive architecture of the de-raining network is searched automatically through a gradient-based search algorithm, which avoids the daunting procedure of the manual design to some extent. Moreover, in order to obtain a robust image de-raining model, a practical and effective multi-to- one training strategy is also presented to allow the de-raining network to get sufficient background information from multiple rainy images with the same background scene, and meanwhile, multiple loss functions including external loss, internal loss, architecture regularization loss, and model complexity loss are jointly optimized to achieve robust de-raining performance and controllable model complexity. Extensive experimental results on both synthetic and realistic rainy images, as well as the down-stream vision applications (i.e., objection detection and segmentation) consistently demonstrate the superiority of our proposed method. The code is publicly available at https://github.com/lcai-gz/MANAS .</description><subject>Complexity</subject><subject>Complexity theory</subject><subject>Computer architecture</subject><subject>Controllability</subject><subject>Image de-raining</subject><subject>Image segmentation</subject><subject>Machine learning</subject><subject>Modules</subject><subject>multi-scale attentive neural architecture search</subject><subject>multi-to-one training strategy</subject><subject>Neural networks</subject><subject>Rain</subject><subject>Regularization</subject><subject>Robust control</subject><subject>Search algorithms</subject><subject>Shape</subject><subject>Task analysis</subject><subject>Training</subject><issn>1051-8215</issn><issn>1558-2205</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kEtPwzAQhC0EEqXwB-ASiXOKn3F8jMqrUgsSLVwt42yKS5oU2yni35PSitPuSDO7mg-hS4JHhGB1sxjP3xYjiikdMYqlINkRGhAh8pRSLI77HQuS5pSIU3QWwgpjwnMuB2g26-rogjU1JEWM0ES3hWSyNktIbiF9Ma5xzTJ5gvjd-s-QbJ3pRedNnRTefrgINnYekjmYXp6jk8rUAS4Oc4he7-8W48d0-vwwGRfT1FIlYiqozFUpc6BQClWa9zLLMmwFVNZSXmJsKo4pMRW1LCslA2WYYARz0hcUCtgQXe_vbnz71UGIetV2vulfaiol41wqxXoX3busb0PwUOmNd2vjfzTBeodN_2HTO2z6gK0PXe1DDgD-AypXnEnBfgFHd2ii</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Cai, Lei</creator><creator>Fu, Yuli</creator><creator>Huo, Wanliang</creator><creator>Xiang, Youjun</creator><creator>Zhu, Tao</creator><creator>Zhang, Ying</creator><creator>Zeng, Huanqiang</creator><creator>Zeng, Delu</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-2109-948X</orcidid><orcidid>https://orcid.org/0000-0002-6548-446X</orcidid><orcidid>https://orcid.org/0000-0001-7322-1873</orcidid><orcidid>https://orcid.org/0000-0002-2802-7745</orcidid><orcidid>https://orcid.org/0000-0003-4763-3529</orcidid><orcidid>https://orcid.org/0000-0002-3244-6459</orcidid></search><sort><creationdate>20230201</creationdate><title>Multiscale Attentive Image De-Raining Networks via Neural Architecture Search</title><author>Cai, Lei ; Fu, Yuli ; Huo, Wanliang ; Xiang, Youjun ; Zhu, Tao ; Zhang, Ying ; Zeng, Huanqiang ; Zeng, Delu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c295t-52789d78e2ed59dabd6660c5efcc24d00af4021af2c36d73e9a353104102259e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Complexity</topic><topic>Complexity theory</topic><topic>Computer architecture</topic><topic>Controllability</topic><topic>Image de-raining</topic><topic>Image segmentation</topic><topic>Machine learning</topic><topic>Modules</topic><topic>multi-scale attentive neural architecture search</topic><topic>multi-to-one training strategy</topic><topic>Neural networks</topic><topic>Rain</topic><topic>Regularization</topic><topic>Robust control</topic><topic>Search algorithms</topic><topic>Shape</topic><topic>Task analysis</topic><topic>Training</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Cai, Lei</creatorcontrib><creatorcontrib>Fu, Yuli</creatorcontrib><creatorcontrib>Huo, Wanliang</creatorcontrib><creatorcontrib>Xiang, Youjun</creatorcontrib><creatorcontrib>Zhu, Tao</creatorcontrib><creatorcontrib>Zhang, Ying</creatorcontrib><creatorcontrib>Zeng, Huanqiang</creatorcontrib><creatorcontrib>Zeng, Delu</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE transactions on circuits and systems for video technology</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Cai, Lei</au><au>Fu, Yuli</au><au>Huo, Wanliang</au><au>Xiang, Youjun</au><au>Zhu, Tao</au><au>Zhang, Ying</au><au>Zeng, Huanqiang</au><au>Zeng, Delu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multiscale Attentive Image De-Raining Networks via Neural Architecture Search</atitle><jtitle>IEEE transactions on circuits and systems for video technology</jtitle><stitle>TCSVT</stitle><date>2023-02-01</date><risdate>2023</risdate><volume>33</volume><issue>2</issue><spage>618</spage><epage>633</epage><pages>618-633</pages><issn>1051-8215</issn><eissn>1558-2205</eissn><coden>ITCTEM</coden><abstract>Multi-scale architectures and attention modules have shown effectiveness in many deep learning-based image de-raining methods. However, manually designing and integrating these two components into a neural network requires a bulk of labor and extensive expertise. In this article, a high-performance multi-scale attentive neural architecture search (MANAS) framework is technically developed for image de-raining. The proposed method formulates a new multi-scale attention search space with multiple flexible modules that are favorite to the image de-raining task. Under the search space, multi-scale attentive cells are built, which are further used to construct a powerful image de-raining network. The internal multi-scale attentive architecture of the de-raining network is searched automatically through a gradient-based search algorithm, which avoids the daunting procedure of the manual design to some extent. Moreover, in order to obtain a robust image de-raining model, a practical and effective multi-to- one training strategy is also presented to allow the de-raining network to get sufficient background information from multiple rainy images with the same background scene, and meanwhile, multiple loss functions including external loss, internal loss, architecture regularization loss, and model complexity loss are jointly optimized to achieve robust de-raining performance and controllable model complexity. Extensive experimental results on both synthetic and realistic rainy images, as well as the down-stream vision applications (i.e., objection detection and segmentation) consistently demonstrate the superiority of our proposed method. The code is publicly available at https://github.com/lcai-gz/MANAS .</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TCSVT.2022.3207516</doi><tpages>16</tpages><orcidid>https://orcid.org/0000-0002-2109-948X</orcidid><orcidid>https://orcid.org/0000-0002-6548-446X</orcidid><orcidid>https://orcid.org/0000-0001-7322-1873</orcidid><orcidid>https://orcid.org/0000-0002-2802-7745</orcidid><orcidid>https://orcid.org/0000-0003-4763-3529</orcidid><orcidid>https://orcid.org/0000-0002-3244-6459</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 1051-8215
ispartof IEEE transactions on circuits and systems for video technology, 2023-02, Vol.33 (2), p.618-633
issn 1051-8215
1558-2205
language eng
recordid cdi_proquest_journals_2773447993
source IEEE Electronic Library (IEL)
subjects Complexity
Complexity theory
Computer architecture
Controllability
Image de-raining
Image segmentation
Machine learning
Modules
multi-scale attentive neural architecture search
multi-to-one training strategy
Neural networks
Rain
Regularization
Robust control
Search algorithms
Shape
Task analysis
Training
title Multiscale Attentive Image De-Raining Networks via Neural Architecture Search
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T19%3A57%3A17IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multiscale%20Attentive%20Image%20De-Raining%20Networks%20via%20Neural%20Architecture%20Search&rft.jtitle=IEEE%20transactions%20on%20circuits%20and%20systems%20for%20video%20technology&rft.au=Cai,%20Lei&rft.date=2023-02-01&rft.volume=33&rft.issue=2&rft.spage=618&rft.epage=633&rft.pages=618-633&rft.issn=1051-8215&rft.eissn=1558-2205&rft.coden=ITCTEM&rft_id=info:doi/10.1109/TCSVT.2022.3207516&rft_dat=%3Cproquest_RIE%3E2773447993%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2773447993&rft_id=info:pmid/&rft_ieee_id=9894375&rfr_iscdi=true