Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks
Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior f...
Gespeichert in:
Veröffentlicht in: | Mobile networks and applications 2021-02, Vol.26 (1), p.40-56 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 56 |
---|---|
container_issue | 1 |
container_start_page | 40 |
container_title | Mobile networks and applications |
container_volume | 26 |
creator | Zhao, Wenyi Yang, Huihua Wang, Jie Pan, Xipeng Cao, Zhiwei |
description | Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called
NEAREST_Gaussian
is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods. |
doi_str_mv | 10.1007/s11036-020-01719-9 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2505799976</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2505799976</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</originalsourceid><addsrcrecordid>eNp9kF1LwzAUhoMoOKd_wKuC19F8NG1zKcPpYH4gKt6FLD3tOrtmJs3Uf29cBe-8OofD874cHoROKTmnhOQXnlLCM0wYwYTmVGK5h0ZU5AwXVPD9uPOC4zSTr4foyPsVIUSIIh2hl0eoG9vhRHdl8tB8QovnsIU2uQ1t3-CpNcEns7WuIZkGH8mkXzob6mUysd3WtqGPN90mdxDcbvQf1r35Y3RQ6dbDye8co-fp1dPkBs_vr2eTyzk2nMoeUxDaZLowUACjpqKLImO5YVxLoWkBlSmMLo2ElJmqFJAuSFqxkuWZ1BwM8DE6G3o3zr4H8L1a2eDiQ14xQUQupcyzSLGBMs5676BSG9estftSlKgff2rwp6I_tfOnZAzxIeQj3NXg_qr_SX0D7S50cQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2505799976</pqid></control><display><type>article</type><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><source>SpringerLink Journals</source><creator>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</creator><creatorcontrib>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</creatorcontrib><description>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called
NEAREST_Gaussian
is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</description><identifier>ISSN: 1383-469X</identifier><identifier>EISSN: 1572-8153</identifier><identifier>DOI: 10.1007/s11036-020-01719-9</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial neural networks ; Communications Engineering ; Computer Communication Networks ; Computer vision ; Depth of field ; Electrical Engineering ; Engineering ; Image classification ; Image processing ; Interpolation ; IT in Business ; Methods ; Networks ; Neural networks ; Object recognition ; Pixels ; Real time ; Visual perception</subject><ispartof>Mobile networks and applications, 2021-02, Vol.26 (1), p.40-56</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC part of Springer Nature 2021</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC part of Springer Nature 2021.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</citedby><cites>FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</cites><orcidid>0000-0002-8339-5081</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11036-020-01719-9$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11036-020-01719-9$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Zhao, Wenyi</creatorcontrib><creatorcontrib>Yang, Huihua</creatorcontrib><creatorcontrib>Wang, Jie</creatorcontrib><creatorcontrib>Pan, Xipeng</creatorcontrib><creatorcontrib>Cao, Zhiwei</creatorcontrib><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><title>Mobile networks and applications</title><addtitle>Mobile Netw Appl</addtitle><description>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called
NEAREST_Gaussian
is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</description><subject>Artificial neural networks</subject><subject>Communications Engineering</subject><subject>Computer Communication Networks</subject><subject>Computer vision</subject><subject>Depth of field</subject><subject>Electrical Engineering</subject><subject>Engineering</subject><subject>Image classification</subject><subject>Image processing</subject><subject>Interpolation</subject><subject>IT in Business</subject><subject>Methods</subject><subject>Networks</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Pixels</subject><subject>Real time</subject><subject>Visual perception</subject><issn>1383-469X</issn><issn>1572-8153</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9kF1LwzAUhoMoOKd_wKuC19F8NG1zKcPpYH4gKt6FLD3tOrtmJs3Uf29cBe-8OofD874cHoROKTmnhOQXnlLCM0wYwYTmVGK5h0ZU5AwXVPD9uPOC4zSTr4foyPsVIUSIIh2hl0eoG9vhRHdl8tB8QovnsIU2uQ1t3-CpNcEns7WuIZkGH8mkXzob6mUysd3WtqGPN90mdxDcbvQf1r35Y3RQ6dbDye8co-fp1dPkBs_vr2eTyzk2nMoeUxDaZLowUACjpqKLImO5YVxLoWkBlSmMLo2ElJmqFJAuSFqxkuWZ1BwM8DE6G3o3zr4H8L1a2eDiQ14xQUQupcyzSLGBMs5676BSG9estftSlKgff2rwp6I_tfOnZAzxIeQj3NXg_qr_SX0D7S50cQ</recordid><startdate>20210201</startdate><enddate>20210201</enddate><creator>Zhao, Wenyi</creator><creator>Yang, Huihua</creator><creator>Wang, Jie</creator><creator>Pan, Xipeng</creator><creator>Cao, Zhiwei</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7SP</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-8339-5081</orcidid></search><sort><creationdate>20210201</creationdate><title>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</title><author>Zhao, Wenyi ; Yang, Huihua ; Wang, Jie ; Pan, Xipeng ; Cao, Zhiwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-1e5ac6a8ce8e21cf1b8627c23a95a18efc8cadc9e42cfd5e4b04f2d2769a3ece3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Artificial neural networks</topic><topic>Communications Engineering</topic><topic>Computer Communication Networks</topic><topic>Computer vision</topic><topic>Depth of field</topic><topic>Electrical Engineering</topic><topic>Engineering</topic><topic>Image classification</topic><topic>Image processing</topic><topic>Interpolation</topic><topic>IT in Business</topic><topic>Methods</topic><topic>Networks</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Pixels</topic><topic>Real time</topic><topic>Visual perception</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhao, Wenyi</creatorcontrib><creatorcontrib>Yang, Huihua</creatorcontrib><creatorcontrib>Wang, Jie</creatorcontrib><creatorcontrib>Pan, Xipeng</creatorcontrib><creatorcontrib>Cao, Zhiwei</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>ProQuest Central Basic</collection><jtitle>Mobile networks and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhao, Wenyi</au><au>Yang, Huihua</au><au>Wang, Jie</au><au>Pan, Xipeng</au><au>Cao, Zhiwei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks</atitle><jtitle>Mobile networks and applications</jtitle><stitle>Mobile Netw Appl</stitle><date>2021-02-01</date><risdate>2021</risdate><volume>26</volume><issue>1</issue><spage>40</spage><epage>56</epage><pages>40-56</pages><issn>1383-469X</issn><eissn>1572-8153</eissn><abstract>Capturing all-in-focus images with 3D scenes is typically a challenging task due to depth of field limitations, and various multi-focus image fusion methods have been employed to generate all-in-focus images. However, existing methods have difficulty simultaneously achieving real-time and superior fusion performance. In this paper, we propose a region- and pixel-based method that can recognize the focus and defocus regions or pixels by the neighborhood information in the source images. The proposed method can obtain satisfactory fusion results and achieve improved real-time performance. First, a convolutional neural network (CNN)-based classifier generates a coarse region-based trimap quickly, which contains focus, defocus and boundary regions. Then, precise fine-tuning is implemented at the boundary regions to address the boundary pixels that are difficult to discriminate by existing methods. Based on a public database, a high-quality dataset is constructed that provides abundant precise pixel-level labels, so that the proposed method can accurately classify the regions and pixels without artifacts. Furthermore, an image interpolation method called
NEAREST_Gaussian
is proposed to improve the recognition ability at the boundary. Experimental results show that the proposed method outperforms other state-of-the-art methods in visual perception and object metrics. Additionally, the proposed method has 80% improved to the conventional CNN-based methods.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11036-020-01719-9</doi><tpages>17</tpages><orcidid>https://orcid.org/0000-0002-8339-5081</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1383-469X |
ispartof | Mobile networks and applications, 2021-02, Vol.26 (1), p.40-56 |
issn | 1383-469X 1572-8153 |
language | eng |
recordid | cdi_proquest_journals_2505799976 |
source | SpringerLink Journals |
subjects | Artificial neural networks Communications Engineering Computer Communication Networks Computer vision Depth of field Electrical Engineering Engineering Image classification Image processing Interpolation IT in Business Methods Networks Neural networks Object recognition Pixels Real time Visual perception |
title | Region- and Pixel-Level Multi-Focus Image Fusion through Convolutional Neural Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T08%3A27%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Region-%20and%20Pixel-Level%20Multi-Focus%20Image%20Fusion%20through%20Convolutional%20Neural%20Networks&rft.jtitle=Mobile%20networks%20and%20applications&rft.au=Zhao,%20Wenyi&rft.date=2021-02-01&rft.volume=26&rft.issue=1&rft.spage=40&rft.epage=56&rft.pages=40-56&rft.issn=1383-469X&rft.eissn=1572-8153&rft_id=info:doi/10.1007/s11036-020-01719-9&rft_dat=%3Cproquest_cross%3E2505799976%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2505799976&rft_id=info:pmid/&rfr_iscdi=true |