Visible and Infrared Image Fusion Using Encoder-Decoder Network
The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The p...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-12 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Ferhat Can Ataman Akar, Gözde Bozdaği |
description | The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}. |
doi_str_mv | 10.48550/arxiv.2412.08073 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2412_08073</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3143450832</sourcerecordid><originalsourceid>FETCH-LOGICAL-a522-75ec825b667858c7070c9b60a74faf2931d98462e2461f163a1d061bc3fd71163</originalsourceid><addsrcrecordid>eNotj01Lw0AYhBdBsNT-AE8GPKe---5nTiK11ULRS_UaNslu2dpu6m7jx783pp5mBoZhHkKuKEy5FgJuTfz2n1PkFKegQbEzMkLGaK454gWZpLQFAJQKhWAjcvfmk692NjOhyZbBRRNtb_ZmY7NFl3wbstfkwyabh7ptbMwf7KDZsz1-tfH9kpw7s0t28q9jsl7M17OnfPXyuJzdr3IjEHMlbK1RVFIqLXStQEFdVBKM4s44LBhtCs0lWuSSOiqZoQ1IWtXMNYr2eUyuT7MDXXmIfm_iT_lHWQ6UfePm1DjE9qOz6Vhu2y6G_lPJKGdcgGbIfgGKglKf</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3143450832</pqid></control><display><type>article</type><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Ferhat Can Ataman ; Akar, Gözde Bozdaği</creator><creatorcontrib>Ferhat Can Ataman ; Akar, Gözde Bozdaği</creatorcontrib><description>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2412.08073</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Computer Science - Computer Vision and Pattern Recognition ; Computer Science - Learning ; Computer vision ; Encoders-Decoders ; Image quality ; Infrared analysis ; Infrared imagery ; Real time ; Visible spectrum</subject><ispartof>arXiv.org, 2024-12</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,780,881,27902</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2412.08073$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1109/ICIP42928.2021.9506740$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Ferhat Can Ataman</creatorcontrib><creatorcontrib>Akar, Gözde Bozdaği</creatorcontrib><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><title>arXiv.org</title><description>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</description><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Computer Science - Learning</subject><subject>Computer vision</subject><subject>Encoders-Decoders</subject><subject>Image quality</subject><subject>Infrared analysis</subject><subject>Infrared imagery</subject><subject>Real time</subject><subject>Visible spectrum</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><sourceid>GOX</sourceid><recordid>eNotj01Lw0AYhBdBsNT-AE8GPKe---5nTiK11ULRS_UaNslu2dpu6m7jx783pp5mBoZhHkKuKEy5FgJuTfz2n1PkFKegQbEzMkLGaK454gWZpLQFAJQKhWAjcvfmk692NjOhyZbBRRNtb_ZmY7NFl3wbstfkwyabh7ptbMwf7KDZsz1-tfH9kpw7s0t28q9jsl7M17OnfPXyuJzdr3IjEHMlbK1RVFIqLXStQEFdVBKM4s44LBhtCs0lWuSSOiqZoQ1IWtXMNYr2eUyuT7MDXXmIfm_iT_lHWQ6UfePm1DjE9qOz6Vhu2y6G_lPJKGdcgGbIfgGKglKf</recordid><startdate>20241211</startdate><enddate>20241211</enddate><creator>Ferhat Can Ataman</creator><creator>Akar, Gözde Bozdaği</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241211</creationdate><title>Visible and Infrared Image Fusion Using Encoder-Decoder Network</title><author>Ferhat Can Ataman ; Akar, Gözde Bozdaği</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a522-75ec825b667858c7070c9b60a74faf2931d98462e2461f163a1d061bc3fd71163</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Computer Science - Learning</topic><topic>Computer vision</topic><topic>Encoders-Decoders</topic><topic>Image quality</topic><topic>Infrared analysis</topic><topic>Infrared imagery</topic><topic>Real time</topic><topic>Visible spectrum</topic><toplevel>online_resources</toplevel><creatorcontrib>Ferhat Can Ataman</creatorcontrib><creatorcontrib>Akar, Gözde Bozdaği</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Ferhat Can Ataman</au><au>Akar, Gözde Bozdaği</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Visible and Infrared Image Fusion Using Encoder-Decoder Network</atitle><jtitle>arXiv.org</jtitle><date>2024-12-11</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>The aim of multispectral image fusion is to combine object or scene features of images with different spectral characteristics to increase the perceptual quality. In this paper, we present a novel learning-based solution to image fusion problem focusing on infrared and visible spectrum images. The proposed solution utilizes only convolution and pooling layers together with a loss function using no-reference quality metrics. The analysis is performed qualitatively and quantitatively on various datasets. The results show better performance than state-of-the-art methods. Also, the size of our network enables real-time performance on embedded devices. Project codes can be found at \url{https://github.com/ferhatcan/pyFusionSR}.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2412.08073</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-12 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2412_08073 |
source | arXiv.org; Free E- Journals |
subjects | Computer Science - Computer Vision and Pattern Recognition Computer Science - Learning Computer vision Encoders-Decoders Image quality Infrared analysis Infrared imagery Real time Visible spectrum |
title | Visible and Infrared Image Fusion Using Encoder-Decoder Network |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T17%3A34%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Visible%20and%20Infrared%20Image%20Fusion%20Using%20Encoder-Decoder%20Network&rft.jtitle=arXiv.org&rft.au=Ferhat%20Can%20Ataman&rft.date=2024-12-11&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2412.08073&rft_dat=%3Cproquest_arxiv%3E3143450832%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3143450832&rft_id=info:pmid/&rfr_iscdi=true |