DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation
Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmen...
Gespeichert in:
Veröffentlicht in: | IEEE transactions on geoscience and remote sensing 2022, Vol.60, p.1-15 |
---|---|
Hauptverfasser: | , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 15 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE transactions on geoscience and remote sensing |
container_volume | 60 |
creator | Kang, Jian Wang, Zhirui Zhu, Ruoxin Xia, Junshi Sun, Xian Fernandez-Beltran, Ruben Plaza, Antonio |
description | Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet . |
doi_str_mv | 10.1109/TGRS.2022.3165209 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_journals_2652699739</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9750128</ieee_id><sourcerecordid>2652699739</sourcerecordid><originalsourceid>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</originalsourceid><addsrcrecordid>eNo9kEtLAzEUhYMoWKs_QNwEXE9NMq_EnVZbi8WCrbgcMpObMWVeJiniv3eGKW7ui3POhQ-ha0pmlBJxt1u-b2eMMDYLaRIzIk7QhMYxD0gSRadoQqhIAsYFO0cXzu0JoVFM0wnaPxm36fwb-Hvcj95UlWlKvIVaNt4U-LVpfypQJeCFbWvcS00hK7yqZQkO69biT5D-C2ywahR00JfG48eDqdSYU9b9QXrTNpfoTMvKwdWxT9HH4nk3fwnWm-Vq_rAOCiZCH2gtQsiTlBfAIh2TuCCCcc6V5qGKFKWgmRp2Snka5kCVLHKtc0h4nghZhFN0O-Z2tv0-gPPZvj3Ypn-ZsR5NIkQail5FR1VhW-cs6Kyzppb2N6MkG5BmA9JsQJodkfaem9FjAOBfL9KYUMbDP-6IdEA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2652699739</pqid></control><display><type>article</type><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><source>IEEE Electronic Library (IEL)</source><creator>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</creator><creatorcontrib>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</creatorcontrib><description>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</description><identifier>ISSN: 0196-2892</identifier><identifier>EISSN: 1558-0644</identifier><identifier>DOI: 10.1109/TGRS.2022.3165209</identifier><identifier>CODEN: IGRSD2</identifier><language>eng</language><publisher>New York: IEEE</publisher><subject>Adaptive optics ; Aggregation ; Artificial neural networks ; Building extraction ; Buildings ; Computer architecture ; deep learning ; Disasters ; Distillation ; Image processing ; Image segmentation ; knowledge distillation ; Knowledge management ; Low level ; Methods ; Mimicry ; missing modality ; Neural networks ; Optical imaging ; Optical sensors ; Radar imaging ; Radar polarimetry ; Remote observing ; Remote sensing ; SAR (radar) ; semantic segmentation ; Semantics ; Synthetic aperture radar ; synthetic aperture radar (SAR) ; transfer learning ; Weather</subject><ispartof>IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-15</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2022</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</citedby><cites>FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</cites><orcidid>0000-0002-5586-6536 ; 0000-0001-6284-3044 ; 0000-0003-2877-0384 ; 0000-0002-0038-9816 ; 0000-0003-1374-8416 ; 0000-0002-9613-1659 ; 0000-0003-4552-8223</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9750128$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,4010,27900,27901,27902,54733</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/9750128$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Kang, Jian</creatorcontrib><creatorcontrib>Wang, Zhirui</creatorcontrib><creatorcontrib>Zhu, Ruoxin</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Sun, Xian</creatorcontrib><creatorcontrib>Fernandez-Beltran, Ruben</creatorcontrib><creatorcontrib>Plaza, Antonio</creatorcontrib><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><title>IEEE transactions on geoscience and remote sensing</title><addtitle>TGRS</addtitle><description>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</description><subject>Adaptive optics</subject><subject>Aggregation</subject><subject>Artificial neural networks</subject><subject>Building extraction</subject><subject>Buildings</subject><subject>Computer architecture</subject><subject>deep learning</subject><subject>Disasters</subject><subject>Distillation</subject><subject>Image processing</subject><subject>Image segmentation</subject><subject>knowledge distillation</subject><subject>Knowledge management</subject><subject>Low level</subject><subject>Methods</subject><subject>Mimicry</subject><subject>missing modality</subject><subject>Neural networks</subject><subject>Optical imaging</subject><subject>Optical sensors</subject><subject>Radar imaging</subject><subject>Radar polarimetry</subject><subject>Remote observing</subject><subject>Remote sensing</subject><subject>SAR (radar)</subject><subject>semantic segmentation</subject><subject>Semantics</subject><subject>Synthetic aperture radar</subject><subject>synthetic aperture radar (SAR)</subject><subject>transfer learning</subject><subject>Weather</subject><issn>0196-2892</issn><issn>1558-0644</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNo9kEtLAzEUhYMoWKs_QNwEXE9NMq_EnVZbi8WCrbgcMpObMWVeJiniv3eGKW7ui3POhQ-ha0pmlBJxt1u-b2eMMDYLaRIzIk7QhMYxD0gSRadoQqhIAsYFO0cXzu0JoVFM0wnaPxm36fwb-Hvcj95UlWlKvIVaNt4U-LVpfypQJeCFbWvcS00hK7yqZQkO69biT5D-C2ywahR00JfG48eDqdSYU9b9QXrTNpfoTMvKwdWxT9HH4nk3fwnWm-Vq_rAOCiZCH2gtQsiTlBfAIh2TuCCCcc6V5qGKFKWgmRp2Snka5kCVLHKtc0h4nghZhFN0O-Z2tv0-gPPZvj3Ypn-ZsR5NIkQail5FR1VhW-cs6Kyzppb2N6MkG5BmA9JsQJodkfaem9FjAOBfL9KYUMbDP-6IdEA</recordid><startdate>2022</startdate><enddate>2022</enddate><creator>Kang, Jian</creator><creator>Wang, Zhirui</creator><creator>Zhu, Ruoxin</creator><creator>Xia, Junshi</creator><creator>Sun, Xian</creator><creator>Fernandez-Beltran, Ruben</creator><creator>Plaza, Antonio</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7UA</scope><scope>8FD</scope><scope>C1K</scope><scope>F1W</scope><scope>FR3</scope><scope>H8D</scope><scope>H96</scope><scope>KR7</scope><scope>L.G</scope><scope>L7M</scope><orcidid>https://orcid.org/0000-0002-5586-6536</orcidid><orcidid>https://orcid.org/0000-0001-6284-3044</orcidid><orcidid>https://orcid.org/0000-0003-2877-0384</orcidid><orcidid>https://orcid.org/0000-0002-0038-9816</orcidid><orcidid>https://orcid.org/0000-0003-1374-8416</orcidid><orcidid>https://orcid.org/0000-0002-9613-1659</orcidid><orcidid>https://orcid.org/0000-0003-4552-8223</orcidid></search><sort><creationdate>2022</creationdate><title>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</title><author>Kang, Jian ; Wang, Zhirui ; Zhu, Ruoxin ; Xia, Junshi ; Sun, Xian ; Fernandez-Beltran, Ruben ; Plaza, Antonio</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c293t-ff93eb678ce24f505c092888df83d4d11ef2d888d11873be1dacbffbe68b69ac3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Adaptive optics</topic><topic>Aggregation</topic><topic>Artificial neural networks</topic><topic>Building extraction</topic><topic>Buildings</topic><topic>Computer architecture</topic><topic>deep learning</topic><topic>Disasters</topic><topic>Distillation</topic><topic>Image processing</topic><topic>Image segmentation</topic><topic>knowledge distillation</topic><topic>Knowledge management</topic><topic>Low level</topic><topic>Methods</topic><topic>Mimicry</topic><topic>missing modality</topic><topic>Neural networks</topic><topic>Optical imaging</topic><topic>Optical sensors</topic><topic>Radar imaging</topic><topic>Radar polarimetry</topic><topic>Remote observing</topic><topic>Remote sensing</topic><topic>SAR (radar)</topic><topic>semantic segmentation</topic><topic>Semantics</topic><topic>Synthetic aperture radar</topic><topic>synthetic aperture radar (SAR)</topic><topic>transfer learning</topic><topic>Weather</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kang, Jian</creatorcontrib><creatorcontrib>Wang, Zhirui</creatorcontrib><creatorcontrib>Zhu, Ruoxin</creatorcontrib><creatorcontrib>Xia, Junshi</creatorcontrib><creatorcontrib>Sun, Xian</creatorcontrib><creatorcontrib>Fernandez-Beltran, Ruben</creatorcontrib><creatorcontrib>Plaza, Antonio</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Water Resources Abstracts</collection><collection>Technology Research Database</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ASFA: Aquatic Sciences and Fisheries Abstracts</collection><collection>Engineering Research Database</collection><collection>Aerospace Database</collection><collection>Aquatic Science & Fisheries Abstracts (ASFA) 2: Ocean Technology, Policy & Non-Living Resources</collection><collection>Civil Engineering Abstracts</collection><collection>Aquatic Science & Fisheries Abstracts (ASFA) Professional</collection><collection>Advanced Technologies Database with Aerospace</collection><jtitle>IEEE transactions on geoscience and remote sensing</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Kang, Jian</au><au>Wang, Zhirui</au><au>Zhu, Ruoxin</au><au>Xia, Junshi</au><au>Sun, Xian</au><au>Fernandez-Beltran, Ruben</au><au>Plaza, Antonio</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation</atitle><jtitle>IEEE transactions on geoscience and remote sensing</jtitle><stitle>TGRS</stitle><date>2022</date><risdate>2022</risdate><volume>60</volume><spage>1</spage><epage>15</epage><pages>1-15</pages><issn>0196-2892</issn><eissn>1558-0644</eissn><coden>IGRSD2</coden><abstract>Synthetic aperture radar (SAR) images provide all-weather and all-time capabilities for Earth observation, which becomes highly beneficial in the field of intelligent remote sensing (RS) image interpretation. Due to these advantages, SAR images have been widely exploited in automatic building segmentation tasks under poor weather conditions, especially when disasters happen. However, compared to optical images, the semantics inherent to SAR images are less rich and interpretable due to factors such as speckle noise and imaging geometry. In this scenario, most state-of-the-art methods are focused on designing advanced network architectures or loss functions for building footprint extraction. However, few works have been oriented toward improving segmentation performance through knowledge transfer from optical images. In this article, we propose a novel method based on the DisOptNet network, which can distill the useful semantic knowledge from optical images into a network only trained with SAR data. Specifically, we first analyze the multilevel feature discrepancies between multiple stages of the networks pretrained on the two image modalities. We observe that feature discrepancies start to increase as the encoding stage gradually changes from low level to high level. Based on such observation, we reuse the early stage features and construct parallel convolutional neural network (CNN) branches that are responsible for capturing high-level domain-specific knowledge for each image modality. The optical branch is aimed at mimicking feature generation at the optical pretrained network given the input SAR images. Then, an aggregation module is introduced to calibrate and fuse the features from different modalities while generating the building segments. Extensive experiments were conducted on a large-scale multisensor all-weather building segmentation dataset with state-of-the-art methods used for comparison. Our experimental results validate the effectiveness of DisOptNet , which demonstrates great potential in the task of weather-independent building footprint generation under real scenarios. The codes of this article will be made publicly available at https://github.com/jiankang1991/TGRS_DisOptNet .</abstract><cop>New York</cop><pub>IEEE</pub><doi>10.1109/TGRS.2022.3165209</doi><tpages>15</tpages><orcidid>https://orcid.org/0000-0002-5586-6536</orcidid><orcidid>https://orcid.org/0000-0001-6284-3044</orcidid><orcidid>https://orcid.org/0000-0003-2877-0384</orcidid><orcidid>https://orcid.org/0000-0002-0038-9816</orcidid><orcidid>https://orcid.org/0000-0003-1374-8416</orcidid><orcidid>https://orcid.org/0000-0002-9613-1659</orcidid><orcidid>https://orcid.org/0000-0003-4552-8223</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 0196-2892 |
ispartof | IEEE transactions on geoscience and remote sensing, 2022, Vol.60, p.1-15 |
issn | 0196-2892 1558-0644 |
language | eng |
recordid | cdi_proquest_journals_2652699739 |
source | IEEE Electronic Library (IEL) |
subjects | Adaptive optics Aggregation Artificial neural networks Building extraction Buildings Computer architecture deep learning Disasters Distillation Image processing Image segmentation knowledge distillation Knowledge management Low level Methods Mimicry missing modality Neural networks Optical imaging Optical sensors Radar imaging Radar polarimetry Remote observing Remote sensing SAR (radar) semantic segmentation Semantics Synthetic aperture radar synthetic aperture radar (SAR) transfer learning Weather |
title | DisOptNet: Distilling Semantic Knowledge From Optical Images for Weather-Independent Building Segmentation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-08T22%3A12%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=DisOptNet:%20Distilling%20Semantic%20Knowledge%20From%20Optical%20Images%20for%20Weather-Independent%20Building%20Segmentation&rft.jtitle=IEEE%20transactions%20on%20geoscience%20and%20remote%20sensing&rft.au=Kang,%20Jian&rft.date=2022&rft.volume=60&rft.spage=1&rft.epage=15&rft.pages=1-15&rft.issn=0196-2892&rft.eissn=1558-0644&rft.coden=IGRSD2&rft_id=info:doi/10.1109/TGRS.2022.3165209&rft_dat=%3Cproquest_RIE%3E2652699739%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2652699739&rft_id=info:pmid/&rft_ieee_id=9750128&rfr_iscdi=true |