A Tri-Attention fusion guided multi-modal segmentation network
•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•...
Gespeichert in:
Veröffentlicht in: | Pattern recognition 2022-04, Vol.124, p.108417, Article 108417 |
---|---|
Hauptverfasser: | , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | 108417 |
container_title | Pattern recognition |
container_volume | 124 |
creator | Zhou, Tongxue Ruan, Su Vera, Pierre Canu, Stéphane |
description | •A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed.
In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method. |
doi_str_mv | 10.1016/j.patcog.2021.108417 |
format | Article |
fullrecord | <record><control><sourceid>hal_cross</sourceid><recordid>TN_cdi_hal_primary_oai_HAL_hal_03710268v1</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S0031320321005938</els_id><sourcerecordid>oai_HAL_hal_03710268v1</sourcerecordid><originalsourceid>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</originalsourceid><addsrcrecordid>eNp9kE9LAzEUxIMoWKvfwMNePWzN22Q36aWwFG2Fgpd6Dmnytqbun5KkFb-9u6549DQw7zcDbwi5BzoDCsXjYXbU0XT7WUYz6C3JQVyQCUjB0hx4dkkmlDJIWUbZNbkJ4UApiP4wIYsy2XqXljFiG13XJtUpDLI_OYs2aU51dGnTWV0nAfdND-kfrMX42fmPW3JV6Trg3a9Oydvz03a5Tjevq5dluUkNk0VMC9QS8oJzShG1yHacmyrfCS6p0DA3xhoz50VhJKdMil01R4Z5JoTlFhAqNiUPY--7rtXRu0b7L9Vpp9blRg0eZQJoVsgz9CwfWeO7EDxWfwGgathLHdS4lxr2UuNefWwxxrD_4-zQq2Actgat82iisp37v-AbKIt0wA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Tri-Attention fusion guided multi-modal segmentation network</title><source>ScienceDirect Journals (5 years ago - present)</source><creator>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</creator><creatorcontrib>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</creatorcontrib><description>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed.
In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</description><identifier>ISSN: 0031-3203</identifier><identifier>EISSN: 1873-5142</identifier><identifier>DOI: 10.1016/j.patcog.2021.108417</identifier><language>eng</language><publisher>Elsevier Ltd</publisher><subject>Brain tumor segmentation ; Correlation ; Deep learning ; Life Sciences ; Multi-modality fusion</subject><ispartof>Pattern recognition, 2022-04, Vol.124, p.108417, Article 108417</ispartof><rights>2021 Elsevier Ltd</rights><rights>Attribution - NonCommercial</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</citedby><cites>FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://dx.doi.org/10.1016/j.patcog.2021.108417$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,780,784,885,3548,27922,27923,45993</link.rule.ids><backlink>$$Uhttps://hal.science/hal-03710268$$DView record in HAL$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhou, Tongxue</creatorcontrib><creatorcontrib>Ruan, Su</creatorcontrib><creatorcontrib>Vera, Pierre</creatorcontrib><creatorcontrib>Canu, Stéphane</creatorcontrib><title>A Tri-Attention fusion guided multi-modal segmentation network</title><title>Pattern recognition</title><description>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed.
In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</description><subject>Brain tumor segmentation</subject><subject>Correlation</subject><subject>Deep learning</subject><subject>Life Sciences</subject><subject>Multi-modality fusion</subject><issn>0031-3203</issn><issn>1873-5142</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp9kE9LAzEUxIMoWKvfwMNePWzN22Q36aWwFG2Fgpd6Dmnytqbun5KkFb-9u6549DQw7zcDbwi5BzoDCsXjYXbU0XT7WUYz6C3JQVyQCUjB0hx4dkkmlDJIWUbZNbkJ4UApiP4wIYsy2XqXljFiG13XJtUpDLI_OYs2aU51dGnTWV0nAfdND-kfrMX42fmPW3JV6Trg3a9Oydvz03a5Tjevq5dluUkNk0VMC9QS8oJzShG1yHacmyrfCS6p0DA3xhoz50VhJKdMil01R4Z5JoTlFhAqNiUPY--7rtXRu0b7L9Vpp9blRg0eZQJoVsgz9CwfWeO7EDxWfwGgathLHdS4lxr2UuNefWwxxrD_4-zQq2Actgat82iisp37v-AbKIt0wA</recordid><startdate>202204</startdate><enddate>202204</enddate><creator>Zhou, Tongxue</creator><creator>Ruan, Su</creator><creator>Vera, Pierre</creator><creator>Canu, Stéphane</creator><general>Elsevier Ltd</general><general>Elsevier</general><scope>AAYXX</scope><scope>CITATION</scope><scope>1XC</scope><scope>VOOES</scope></search><sort><creationdate>202204</creationdate><title>A Tri-Attention fusion guided multi-modal segmentation network</title><author>Zhou, Tongxue ; Ruan, Su ; Vera, Pierre ; Canu, Stéphane</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c386t-6ea81564400eea72b44cf5b74807a19ccdcc9466c840387bf9e3e5277d4d1e1f3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Brain tumor segmentation</topic><topic>Correlation</topic><topic>Deep learning</topic><topic>Life Sciences</topic><topic>Multi-modality fusion</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Tongxue</creatorcontrib><creatorcontrib>Ruan, Su</creatorcontrib><creatorcontrib>Vera, Pierre</creatorcontrib><creatorcontrib>Canu, Stéphane</creatorcontrib><collection>CrossRef</collection><collection>Hyper Article en Ligne (HAL)</collection><collection>Hyper Article en Ligne (HAL) (Open Access)</collection><jtitle>Pattern recognition</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhou, Tongxue</au><au>Ruan, Su</au><au>Vera, Pierre</au><au>Canu, Stéphane</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Tri-Attention fusion guided multi-modal segmentation network</atitle><jtitle>Pattern recognition</jtitle><date>2022-04</date><risdate>2022</risdate><volume>124</volume><spage>108417</spage><pages>108417-</pages><artnum>108417</artnum><issn>0031-3203</issn><eissn>1873-5142</eissn><abstract>•A novel correlation description block is introduced to discover the latent multi-source correlation between modalities.•A constraint based on the correlation using KL divergence is proposed to aide the segmentation network to extract the correlated feature representation for a better segmentation.•A tri-attention fusion strategy is proposed to recalibrate the feature representation along modality-attention, spatial-attention and correlation-attention paths.•The first 3D multi-modal brain tumor segmentation network guided by tri-attention fusion is proposed.
In the field of multimodal segmentation, the correlation between different modalities can be considered for improving the segmentation results. Considering the correlation between different MR modalities, in this paper, we propose a multi-modality segmentation network guided by a novel tri-attention fusion. Our network includes N model-independent encoding paths with N image sources, a tri-attention fusion block, a dual-attention fusion block, and a decoding path. The model independent encoding paths can capture modality-specific features from the N modalities. Considering that not all the features extracted from the encoders are useful for segmentation, we propose to use dual attention based fusion to re-weight the features along the modality and space paths, which can suppress less informative features and emphasize the useful ones for each modality at different positions. Since there exists a strong correlation between different modalities, based on the dual attention fusion block, we propose a correlation attention module to form the tri-attention fusion block. In the correlation attention module, a correlation description block is first used to learn the correlation between modalities and then a constraint based on the correlation is used to guide the network to learn the latent correlated features which are more relevant for segmentation. Finally, the obtained fused feature representation is projected by the decoder to obtain the segmentation results. Our experiment results tested on BraTS 2018 dataset for brain tumor segmentation demonstrate the effectiveness of our proposed method.</abstract><pub>Elsevier Ltd</pub><doi>10.1016/j.patcog.2021.108417</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0031-3203 |
ispartof | Pattern recognition, 2022-04, Vol.124, p.108417, Article 108417 |
issn | 0031-3203 1873-5142 |
language | eng |
recordid | cdi_hal_primary_oai_HAL_hal_03710268v1 |
source | ScienceDirect Journals (5 years ago - present) |
subjects | Brain tumor segmentation Correlation Deep learning Life Sciences Multi-modality fusion |
title | A Tri-Attention fusion guided multi-modal segmentation network |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T22%3A18%3A03IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-hal_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Tri-Attention%20fusion%20guided%20multi-modal%20segmentation%20network&rft.jtitle=Pattern%20recognition&rft.au=Zhou,%20Tongxue&rft.date=2022-04&rft.volume=124&rft.spage=108417&rft.pages=108417-&rft.artnum=108417&rft.issn=0031-3203&rft.eissn=1873-5142&rft_id=info:doi/10.1016/j.patcog.2021.108417&rft_dat=%3Chal_cross%3Eoai_HAL_hal_03710268v1%3C/hal_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_els_id=S0031320321005938&rfr_iscdi=true |