Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks

Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior f...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Mobile networks and applications 2021-02, Vol.26 (1), p.40-56
Hauptverfasser: Zhao, Wenyi, Yang, Huihua, Wang, Jie, Pan, Xipeng, Cao, Zhiwei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 56
container_issue 1
container_start_page 40
container_title Mobile networks and applications
container_volume 26
creator Zhao, Wenyi
Yang, Huihua
Wang, Jie
Pan, Xipeng
Cao, Zhiwei
description Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called NEAREST_Gaussian is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.
doi_str_mv 10.1007/s11036-020-01719-9
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2505799976</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2505799976</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</originalsourceid><addsrcrecordid>eNp9kF1LwzAUhoMoOKd_wKuC19F8NG1zKcPpYH4gKt6FLD3tOrtmJs3Uf29cBe-8OofD874cHoROKTmnhOQXnlLCM0wYwYTmVGK5h0ZU5AwXVPD9uPOC4zSTr4foyPsVIUSIIh2hl0eoG9vhRHdl8tB8QovnsIU2uQ1t3-CpNcEns7WuIZkGH8mkXzob6mUysd3WtqGPN90mdxDcbvQf1r35Y3RQ6dbDye8co-fp1dPkBs_vr2eTyzk2nMoeUxDaZLowUACjpqKLImO5YVxLoWkBlSmMLo2ElJmqFJAuSFqxkuWZ1BwM8DE6G3o3zr4H8L1a2eDiQ14xQUQupcyzSLGBMs5676BSG9estftSlKgff2rwp6I_tfOnZAzxIeQj3NXg_qr_SX0D7S50cQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2505799976</pqid></control><display><type>article</type><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><source>SpringerLink Journals</source><creator>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</creator><creatorcontrib>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</creatorcontrib><description>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called NEAREST_Gaussian is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</description><identifier>ISSN: 1383-469X</identifier><identifier>EISSN: 1572-8153</identifier><identifier>DOI: 10.1007/s11036-020-01719-9</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial neural networks ; Communications Engineering ; Computer Communication Networks ; Computer vision ; Depth of field ; Electrical Engineering ; Engineering ; Image classification ; Image processing ; Interpolation ; IT in Business ; Methods ; Networks ; Neural networks ; Object recognition ; Pixels ; Real time ; Visual perception</subject><ispartof>Mobile networks and applications, 2021-02, Vol.26 (1), p.40-56</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC part of Springer Nature 2021</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC part of Springer Nature 2021.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</citedby><cites>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</cites><orcidid>0000-0002-8339-5081</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11036-020-01719-9$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11036-020-01719-9$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Zhao, Wenyi</creatorcontrib><creatorcontrib>Yang, Huihua</creatorcontrib><creatorcontrib>Wang, Jie</creatorcontrib><creatorcontrib>Pan, Xipeng</creatorcontrib><creatorcontrib>Cao, Zhiwei</creatorcontrib><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><title>Mobile networks and applications</title><addtitle>Mobile Netw Appl</addtitle><description>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called NEAREST_Gaussian is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</description><subject>Artificial neural networks</subject><subject>Communications Engineering</subject><subject>Computer Communication Networks</subject><subject>Computer vision</subject><subject>Depth of field</subject><subject>Electrical Engineering</subject><subject>Engineering</subject><subject>Image classification</subject><subject>Image processing</subject><subject>Interpolation</subject><subject>IT in Business</subject><subject>Methods</subject><subject>Networks</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Pixels</subject><subject>Real time</subject><subject>Visual perception</subject><issn>1383-469X</issn><issn>1572-8153</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9kF1LwzAUhoMoOKd_wKuC19F8NG1zKcPpYH4gKt6FLD3tOrtmJs3Uf29cBe-8OofD874cHoROKTmnhOQXnlLCM0wYwYTmVGK5h0ZU5AwXVPD9uPOC4zSTr4foyPsVIUSIIh2hl0eoG9vhRHdl8tB8QovnsIU2uQ1t3-CpNcEns7WuIZkGH8mkXzob6mUysd3WtqGPN90mdxDcbvQf1r35Y3RQ6dbDye8co-fp1dPkBs_vr2eTyzk2nMoeUxDaZLowUACjpqKLImO5YVxLoWkBlSmMLo2ElJmqFJAuSFqxkuWZ1BwM8DE6G3o3zr4H8L1a2eDiQ14xQUQupcyzSLGBMs5676BSG9estftSlKgff2rwp6I_tfOnZAzxIeQj3NXg_qr_SX0D7S50cQ</recordid><startdate>20210201</startdate><enddate>20210201</enddate><creator>Zhao, Wenyi</creator><creator>Yang, Huihua</creator><creator>Wang, Jie</creator><creator>Pan, Xipeng</creator><creator>Cao, Zhiwei</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7SP</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-8339-5081</orcidid></search><sort><creationdate>20210201</creationdate><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><author>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Artificial neural networks</topic><topic>Communications Engineering</topic><topic>Computer Communication Networks</topic><topic>Computer vision</topic><topic>Depth of field</topic><topic>Electrical Engineering</topic><topic>Engineering</topic><topic>Image classification</topic><topic>Image processing</topic><topic>Interpolation</topic><topic>IT in Business</topic><topic>Methods</topic><topic>Networks</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Pixels</topic><topic>Real time</topic><topic>Visual perception</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhao, Wenyi</creatorcontrib><creatorcontrib>Yang, Huihua</creatorcontrib><creatorcontrib>Wang, Jie</creatorcontrib><creatorcontrib>Pan, Xipeng</creatorcontrib><creatorcontrib>Cao, Zhiwei</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Mobile networks and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhao, Wenyi</au><au>Yang, Huihua</au><au>Wang, Jie</au><au>Pan, Xipeng</au><au>Cao, Zhiwei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</atitle><jtitle>Mobile networks and applications</jtitle><stitle>Mobile Netw Appl</stitle><date>2021-02-01</date><risdate>2021</risdate><volume>26</volume><issue>1</issue><spage>40</spage><epage>56</epage><pages>40-56</pages><issn>1383-469X</issn><eissn>1572-8153</eissn><abstract>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called NEAREST_Gaussian is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11036-020-01719-9</doi><tpages>17</tpages><orcidid>https://orcid.org/0000-0002-8339-5081</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 1383-469X
ispartof Mobile networks and applications, 2021-02, Vol.26 (1), p.40-56
issn 1383-469X
1572-8153
language eng
recordid cdi_proquest_journals_2505799976
source SpringerLink Journals
subjects Artificial neural networks
Communications Engineering
Computer Communication Networks
Computer vision
Depth of field
Electrical Engineering
Engineering
Image classification
Image processing
Interpolation
IT in Business
Methods
Networks
Neural networks
Object recognition
Pixels
Real time
Visual perception
title Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T08%3A27%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Region-%20and%20Pixel-Level%20Multi-Focus%20Image%20Fusion%20through%20Convolutional%20Neural%20Networks&rft.jtitle=Mobile%20networks%20and%20applications&rft.au=Zhao,%20Wenyi&rft.date=2021-02-01&rft.volume=26&rft.issue=1&rft.spage=40&rft.epage=56&rft.pages=40-56&rft.issn=1383-469X&rft.eissn=1572-8153&rft_id=info:doi/10.1007/s11036-020-01719-9&rft_dat=%3Cproquest_cross%3E2505799976%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2505799976&rft_id=info:pmid/&rfr_iscdi=true