A Tri-Attention fusion guided multi-modal segmentation network

•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Pattern recognition 2022-04, Vol.124, p.108417, Article 108417
Hauptverfasser: Zhou, Tongxue, Ruan, Su, Vera, Pierre, Canu, Stéphane
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page 108417
container_title Pattern recognition
container_volume 124
creator Zhou, Tongxue
Ruan, Su
Vera, Pierre
Canu, Stéphane
description •A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed. In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.
doi_str_mv 10.1016/j.patcog.2021.108417
format Article
fullrecord <record><control><sourceid>hal_cross</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_03710268v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0031320321005938</els_id><sourcerecordid>oai_HAL_hal_03710268v1</sourcerecordid><originalsourceid>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</originalsourceid><addsrcrecordid>eNp9kE9LAzEUxIMoWKvfwMNePWzN22Q36aWwFG2Fgpd6Dmnytqbun5KkFb-9u6549DQw7zcDbwi5BzoDCsXjYXbU0XT7WUYz6C3JQVyQCUjB0hx4dkkmlDJIWUbZNbkJ4UApiP4wIYsy2XqXljFiG13XJtUpDLI_OYs2aU51dGnTWV0nAfdND-kfrMX42fmPW3JV6Trg3a9Oydvz03a5Tjevq5dluUkNk0VMC9QS8oJzShG1yHacmyrfCS6p0DA3xhoz50VhJKdMil01R4Z5JoTlFhAqNiUPY--7rtXRu0b7L9Vpp9blRg0eZQJoVsgz9CwfWeO7EDxWfwGgathLHdS4lxr2UuNefWwxxrD_4-zQq2Actgat82iisp37v-AbKIt0wA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Tri-Attention fusion guided multi-modal segmentation network</title><source>ScienceDirect Journals (5 years ago - present)</source><creator>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</creator><creatorcontrib>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</creatorcontrib><description>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed. In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</description><identifier>ISSN: 0031-3203</identifier><identifier>EISSN: 1873-5142</identifier><identifier>DOI: 10.1016/j.patcog.2021.108417</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Brain tumor segmentation ; Correlation ; Deep learning ; Life Sciences ; Multi-modality fusion</subject><ispartof>Pattern recognition, 2022-04, Vol.124, p.108417, Article 108417</ispartof><rights>2021 Elsevier Ltd</rights><rights>Attribution - NonCommercial</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</citedby><cites>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.patcog.2021.108417$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,780,784,885,3548,27922,27923,45993</link.rule.ids><backlink>$$Uhttps://hal.science/hal-03710268$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhou, Tongxue</creatorcontrib><creatorcontrib>Ruan, Su</creatorcontrib><creatorcontrib>Vera, Pierre</creatorcontrib><creatorcontrib>Canu, Stéphane</creatorcontrib><title>A Tri-Attention fusion guided multi-modal segmentation network</title><title>Pattern recognition</title><description>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed. In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</description><subject>Brain tumor segmentation</subject><subject>Correlation</subject><subject>Deep learning</subject><subject>Life Sciences</subject><subject>Multi-modality fusion</subject><issn>0031-3203</issn><issn>1873-5142</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE9LAzEUxIMoWKvfwMNePWzN22Q36aWwFG2Fgpd6Dmnytqbun5KkFb-9u6549DQw7zcDbwi5BzoDCsXjYXbU0XT7WUYz6C3JQVyQCUjB0hx4dkkmlDJIWUbZNbkJ4UApiP4wIYsy2XqXljFiG13XJtUpDLI_OYs2aU51dGnTWV0nAfdND-kfrMX42fmPW3JV6Trg3a9Oydvz03a5Tjevq5dluUkNk0VMC9QS8oJzShG1yHacmyrfCS6p0DA3xhoz50VhJKdMil01R4Z5JoTlFhAqNiUPY--7rtXRu0b7L9Vpp9blRg0eZQJoVsgz9CwfWeO7EDxWfwGgathLHdS4lxr2UuNefWwxxrD_4-zQq2Actgat82iisp37v-AbKIt0wA</recordid><startdate>202204</startdate><enddate>202204</enddate><creator>Zhou, Tongxue</creator><creator>Ruan, Su</creator><creator>Vera, Pierre</creator><creator>Canu, Stéphane</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>1XC</scope><scope>VOOES</scope></search><sort><creationdate>202204</creationdate><title>A Tri-Attention fusion guided multi-modal segmentation network</title><author>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Brain tumor segmentation</topic><topic>Correlation</topic><topic>Deep learning</topic><topic>Life Sciences</topic><topic>Multi-modality fusion</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Tongxue</creatorcontrib><creatorcontrib>Ruan, Su</creatorcontrib><creatorcontrib>Vera, Pierre</creatorcontrib><creatorcontrib>Canu, Stéphane</creatorcontrib><collection>CrossRef</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>Pattern recognition</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhou, Tongxue</au><au>Ruan, Su</au><au>Vera, Pierre</au><au>Canu, Stéphane</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Tri-Attention fusion guided multi-modal segmentation network</atitle><jtitle>Pattern recognition</jtitle><date>2022-04</date><risdate>2022</risdate><volume>124</volume><spage>108417</spage><pages>108417-</pages><artnum>108417</artnum><issn>0031-3203</issn><eissn>1873-5142</eissn><abstract>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed. In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.patcog.2021.108417</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0031-3203
ispartof Pattern recognition, 2022-04, Vol.124, p.108417, Article 108417
issn 0031-3203
1873-5142
language eng
recordid cdi_hal_primary_oai_HAL_hal_03710268v1
source ScienceDirect Journals (5 years ago - present)
subjects Brain tumor segmentation
Correlation
Deep learning
Life Sciences
Multi-modality fusion
title A Tri-Attention fusion guided multi-modal segmentation network
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T22%3A18%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-hal_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Tri-Attention%20fusion%20guided%20multi-modal%20segmentation%20network&rft.jtitle=Pattern%20recognition&rft.au=Zhou,%20Tongxue&rft.date=2022-04&rft.volume=124&rft.spage=108417&rft.pages=108417-&rft.artnum=108417&rft.issn=0031-3203&rft.eissn=1873-5142&rft_id=info:doi/10.1016/j.patcog.2021.108417&rft_dat=%3Chal_cross%3Eoai_HAL_hal_03710268v1%3C/hal_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_els_id=S0031320321005938&rfr_iscdi=true