DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation

Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmen...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on geoscience and remote sensing 2022, Vol.60, p.1-15
Hauptverfasser: Kang, Jian, Wang, Zhirui, Zhu, Ruoxin, Xia, Junshi, Sun, Xian, Fernandez-Beltran, Ruben, Plaza, Antonio
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 15
container_issue
container_start_page 1
container_title IEEE transactions on geoscience and remote sensing
container_volume 60
creator Kang, Jian
Wang, Zhirui
Zhu, Ruoxin
Xia, Junshi
Sun, Xian
Fernandez-Beltran, Ruben
Plaza, Antonio
description Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .
doi_str_mv 10.1109/TGRS.2022.3165209
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_journals_2652699739</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9750128</ieee_id><sourcerecordid>2652699739</sourcerecordid><originalsourceid>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</originalsourceid><addsrcrecordid>eNo9kEtLAzEUhYMoWKs_QNwEXE9NMq_EnVZbi8WCrbgcMpObMWVeJiniv3eGKW7ui3POhQ-ha0pmlBJxt1u-b2eMMDYLaRIzIk7QhMYxD0gSRadoQqhIAsYFO0cXzu0JoVFM0wnaPxm36fwb-Hvcj95UlWlKvIVaNt4U-LVpfypQJeCFbWvcS00hK7yqZQkO69biT5D-C2ywahR00JfG48eDqdSYU9b9QXrTNpfoTMvKwdWxT9HH4nk3fwnWm-Vq_rAOCiZCH2gtQsiTlBfAIh2TuCCCcc6V5qGKFKWgmRp2Snka5kCVLHKtc0h4nghZhFN0O-Z2tv0-gPPZvj3Ypn-ZsR5NIkQail5FR1VhW-cs6Kyzppb2N6MkG5BmA9JsQJodkfaem9FjAOBfL9KYUMbDP-6IdEA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2652699739</pqid></control><display><type>article</type><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><source>IEEE Electronic Library (IEL)</source><creator>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</creator><creatorcontrib>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</creatorcontrib><description>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</description><identifier>ISSN: 0196-2892</identifier><identifier>EISSN: 1558-0644</identifier><identifier>DOI: 10.1109/TGRS.2022.3165209</identifier><identifier>CODEN: IGRSD2</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Adaptive optics ; Aggregation ; Artificial neural networks ; Building extraction ; Buildings ; Computer architecture ; deep learning ; Disasters ; Distillation ; Image processing ; Image segmentation ; knowledge distillation ; Knowledge management ; Low level ; Methods ; Mimicry ; missing modality ; Neural networks ; Optical imaging ; Optical sensors ; Radar imaging ; Radar polarimetry ; Remote observing ; Remote sensing ; SAR (radar) ; semantic segmentation ; Semantics ; Synthetic aperture radar ; synthetic aperture radar (SAR) ; transfer learning ; Weather</subject><ispartof>IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-15</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</citedby><cites>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</cites><orcidid>0000-0002-5586-6536 ; 0000-0001-6284-3044 ; 0000-0003-2877-0384 ; 0000-0002-0038-9816 ; 0000-0003-1374-8416 ; 0000-0002-9613-1659 ; 0000-0003-4552-8223</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9750128$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,4010,27900,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9750128$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Kang, Jian</creatorcontrib><creatorcontrib>Wang, Zhirui</creatorcontrib><creatorcontrib>Zhu, Ruoxin</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Sun, Xian</creatorcontrib><creatorcontrib>Fernandez-Beltran, Ruben</creatorcontrib><creatorcontrib>Plaza, Antonio</creatorcontrib><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><title>IEEE transactions on geoscience and remote sensing</title><addtitle>TGRS</addtitle><description>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</description><subject>Adaptive optics</subject><subject>Aggregation</subject><subject>Artificial neural networks</subject><subject>Building extraction</subject><subject>Buildings</subject><subject>Computer architecture</subject><subject>deep learning</subject><subject>Disasters</subject><subject>Distillation</subject><subject>Image processing</subject><subject>Image segmentation</subject><subject>knowledge distillation</subject><subject>Knowledge management</subject><subject>Low level</subject><subject>Methods</subject><subject>Mimicry</subject><subject>missing modality</subject><subject>Neural networks</subject><subject>Optical imaging</subject><subject>Optical sensors</subject><subject>Radar imaging</subject><subject>Radar polarimetry</subject><subject>Remote observing</subject><subject>Remote sensing</subject><subject>SAR (radar)</subject><subject>semantic segmentation</subject><subject>Semantics</subject><subject>Synthetic aperture radar</subject><subject>synthetic aperture radar (SAR)</subject><subject>transfer learning</subject><subject>Weather</subject><issn>0196-2892</issn><issn>1558-0644</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kEtLAzEUhYMoWKs_QNwEXE9NMq_EnVZbi8WCrbgcMpObMWVeJiniv3eGKW7ui3POhQ-ha0pmlBJxt1u-b2eMMDYLaRIzIk7QhMYxD0gSRadoQqhIAsYFO0cXzu0JoVFM0wnaPxm36fwb-Hvcj95UlWlKvIVaNt4U-LVpfypQJeCFbWvcS00hK7yqZQkO69biT5D-C2ywahR00JfG48eDqdSYU9b9QXrTNpfoTMvKwdWxT9HH4nk3fwnWm-Vq_rAOCiZCH2gtQsiTlBfAIh2TuCCCcc6V5qGKFKWgmRp2Snka5kCVLHKtc0h4nghZhFN0O-Z2tv0-gPPZvj3Ypn-ZsR5NIkQail5FR1VhW-cs6Kyzppb2N6MkG5BmA9JsQJodkfaem9FjAOBfL9KYUMbDP-6IdEA</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Kang, Jian</creator><creator>Wang, Zhirui</creator><creator>Zhu, Ruoxin</creator><creator>Xia, Junshi</creator><creator>Sun, Xian</creator><creator>Fernandez-Beltran, Ruben</creator><creator>Plaza, Antonio</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7UA</scope><scope>8FD</scope><scope>C1K</scope><scope>F1W</scope><scope>FR3</scope><scope>H8D</scope><scope>H96</scope><scope>KR7</scope><scope>L.G</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0002-5586-6536</orcidid><orcidid>https://orcid.org/0000-0001-6284-3044</orcidid><orcidid>https://orcid.org/0000-0003-2877-0384</orcidid><orcidid>https://orcid.org/0000-0002-0038-9816</orcidid><orcidid>https://orcid.org/0000-0003-1374-8416</orcidid><orcidid>https://orcid.org/0000-0002-9613-1659</orcidid><orcidid>https://orcid.org/0000-0003-4552-8223</orcidid></search><sort><creationdate>2022</creationdate><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><author>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Adaptive optics</topic><topic>Aggregation</topic><topic>Artificial neural networks</topic><topic>Building extraction</topic><topic>Buildings</topic><topic>Computer architecture</topic><topic>deep learning</topic><topic>Disasters</topic><topic>Distillation</topic><topic>Image processing</topic><topic>Image segmentation</topic><topic>knowledge distillation</topic><topic>Knowledge management</topic><topic>Low level</topic><topic>Methods</topic><topic>Mimicry</topic><topic>missing modality</topic><topic>Neural networks</topic><topic>Optical imaging</topic><topic>Optical sensors</topic><topic>Radar imaging</topic><topic>Radar polarimetry</topic><topic>Remote observing</topic><topic>Remote sensing</topic><topic>SAR (radar)</topic><topic>semantic segmentation</topic><topic>Semantics</topic><topic>Synthetic aperture radar</topic><topic>synthetic aperture radar (SAR)</topic><topic>transfer learning</topic><topic>Weather</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kang, Jian</creatorcontrib><creatorcontrib>Wang, Zhirui</creatorcontrib><creatorcontrib>Zhu, Ruoxin</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Sun, Xian</creatorcontrib><creatorcontrib>Fernandez-Beltran, Ruben</creatorcontrib><creatorcontrib>Plaza, Antonio</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Water Resources Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ASFA: Aquatic Sciences and Fisheries Abstracts</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Aquatic Science &amp; Fisheries Abstracts (ASFA) 2: Ocean Technology, Policy &amp; Non-Living Resources</collection><collection>Civil Engineering Abstracts</collection><collection>Aquatic Science &amp; Fisheries Abstracts (ASFA) Professional</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE transactions on geoscience and remote sensing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kang, Jian</au><au>Wang, Zhirui</au><au>Zhu, Ruoxin</au><au>Xia, Junshi</au><au>Sun, Xian</au><au>Fernandez-Beltran, Ruben</au><au>Plaza, Antonio</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</atitle><jtitle>IEEE transactions on geoscience and remote sensing</jtitle><stitle>TGRS</stitle><date>2022</date><risdate>2022</risdate><volume>60</volume><spage>1</spage><epage>15</epage><pages>1-15</pages><issn>0196-2892</issn><eissn>1558-0644</eissn><coden>IGRSD2</coden><abstract>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TGRS.2022.3165209</doi><tpages>15</tpages><orcidid>https://orcid.org/0000-0002-5586-6536</orcidid><orcidid>https://orcid.org/0000-0001-6284-3044</orcidid><orcidid>https://orcid.org/0000-0003-2877-0384</orcidid><orcidid>https://orcid.org/0000-0002-0038-9816</orcidid><orcidid>https://orcid.org/0000-0003-1374-8416</orcidid><orcidid>https://orcid.org/0000-0002-9613-1659</orcidid><orcidid>https://orcid.org/0000-0003-4552-8223</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 0196-2892
ispartof IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-15
issn 0196-2892
1558-0644
language eng
recordid cdi_proquest_journals_2652699739
source IEEE Electronic Library (IEL)
subjects Adaptive optics
Aggregation
Artificial neural networks
Building extraction
Buildings
Computer architecture
deep learning
Disasters
Distillation
Image processing
Image segmentation
knowledge distillation
Knowledge management
Low level
Methods
Mimicry
missing modality
Neural networks
Optical imaging
Optical sensors
Radar imaging
Radar polarimetry
Remote observing
Remote sensing
SAR (radar)
semantic segmentation
Semantics
Synthetic aperture radar
synthetic aperture radar (SAR)
transfer learning
Weather
title DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T22%3A12%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DisOptNet:%20Distilling%20Semantic%20Knowledge%20From%20Optical%20Images%20for%20Weather-Independent%20Building%20Segmentation&rft.jtitle=IEEE%20transactions%20on%20geoscience%20and%20remote%20sensing&rft.au=Kang,%20Jian&rft.date=2022&rft.volume=60&rft.spage=1&rft.epage=15&rft.pages=1-15&rft.issn=0196-2892&rft.eissn=1558-0644&rft.coden=IGRSD2&rft_id=info:doi/10.1109/TGRS.2022.3165209&rft_dat=%3Cproquest_RIE%3E2652699739%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2652699739&rft_id=info:pmid/&rft_ieee_id=9750128&rfr_iscdi=true