Cascaded hybrid residual U-Net for glioma segmentation

Glioma segmentation is critical for making surgical plans. Recently, the traditional glioma segmentation method is less competitive with two deep learning segmentation strategies: the patch-based method which focuses more on the local feature for each pixel, and the image-based method which fully le...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Multimedia tools and applications 2020-09, Vol.79 (33-34), p.24929-24947
Hauptverfasser: Long, Jiaosong, Ma, Guangzhi, Liu, Hong, Song, Enmin, Hung, Chih-Cheng, Xu, Xiangyang, Jin, Renchao, Zhuang, Yuzhou, Liu, DaiYang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 24947
container_issue 33-34
container_start_page 24929
container_title Multimedia tools and applications
container_volume 79
creator Long, Jiaosong
Ma, Guangzhi
Liu, Hong
Song, Enmin
Hung, Chih-Cheng
Xu, Xiangyang
Jin, Renchao
Zhuang, Yuzhou
Liu, DaiYang
Ma, Guangzhi
Song, Enmin
description Glioma segmentation is critical for making surgical plans. Recently, the traditional glioma segmentation method is less competitive with two deep learning segmentation strategies: the patch-based method which focuses more on the local feature for each pixel, and the image-based method which fully leverages the global feature and captures the overall shape, size and other characteristics of the lesion in a neighborhood of a pixel. In this study, we investigate and integrate the advantages of 2-D and 3-D image-based architectures, and propose a new convolutional neural network called the Cascaded Hybrid Residual U-Net (CHR-U-Net) for MRI glioma segmentation. The CHR-U-Net exploits both the 2D local features as well as the 3D global spatial contextual information simultaneously. In the first-level of CHR-U-Net, the R-2D-U-Net combines the 2D-U-Net and the residual unit for quick lesion area detecting without any miss. To prevent from missing false-positive pixels, the output of R-2D-U-Net is resampled by using the hard-mining to collect more possible false-positive samples. In the second-level of CHR-U-Net, the axial, coronal, and sagittal 3D-U-Nets are trained to predict whether pixels belong to the area of glioma. The results of three 3D-U-Nets are fused to improve the accuracy and reduce false positives. The database of 2017 BRATS challenge were used in our experiments for the verification. The Dices and Sensitivities of Enhancing, Whole, and Core areas were calculated. The Dices are 0.73, 0.90, and 0.83 and the Sensitivities are 0.83, 0.90, and 0.82, respectively, for the axial, coronal, and sagittal 3D-U-Nets. Experimental results show that the proposed model significantly improves the performance of glioma segmentation.
doi_str_mv 10.1007/s11042-020-09210-z
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2435937552</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2435937552</sourcerecordid><originalsourceid>FETCH-LOGICAL-c319t-935598d0c73096acb6b95b51b83d78859a5eef2bf30243d9e364b4fccc546fdb3</originalsourceid><addsrcrecordid>eNp9kDtPwzAUhS0EEqXwB5giMRuu7TiOR1TxkipY6Gz5GVLlUex0aH89hiCxMd0znO9c6UPomsAtARB3iRAoKQYKGCQlgI8naEG4YFgISk5zZjVgwYGco4uUtgCk4rRcoGqlk9XOu-LjYGLriuhT6_a6Kzb41U9FGGPRdO3Y6yL5pvfDpKd2HC7RWdBd8le_d4k2jw_vq2e8fnt6Wd2vsWVETlgyzmXtwAoGstLWVEZyw4mpmRN1zaXm3gdqAgNaMic9q0pTBmstL6vgDFuim3l3F8fPvU-T2o77OOSXKgNcMsE5zS06t2wcU4o-qF1sex0PioD69qNmPyr7UT9-1DFDbIZSLg-Nj3_T_1Bf2GpnzA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2435937552</pqid></control><display><type>article</type><title>Cascaded hybrid residual U-Net for glioma segmentation</title><source>SpringerLink Journals - AutoHoldings</source><creator>Long, Jiaosong ; Ma, Guangzhi ; Liu, Hong ; Song, Enmin ; Hung, Chih-Cheng ; Xu, Xiangyang ; Jin, Renchao ; Zhuang, Yuzhou ; Liu, DaiYang ; Ma, Guangzhi ; Song, Enmin</creator><creatorcontrib>Long, Jiaosong ; Ma, Guangzhi ; Liu, Hong ; Song, Enmin ; Hung, Chih-Cheng ; Xu, Xiangyang ; Jin, Renchao ; Zhuang, Yuzhou ; Liu, DaiYang ; Ma, Guangzhi ; Song, Enmin</creatorcontrib><description>Glioma segmentation is critical for making surgical plans. Recently, the traditional glioma segmentation method is less competitive with two deep learning segmentation strategies: the patch-based method which focuses more on the local feature for each pixel, and the image-based method which fully leverages the global feature and captures the overall shape, size and other characteristics of the lesion in a neighborhood of a pixel. In this study, we investigate and integrate the advantages of 2-D and 3-D image-based architectures, and propose a new convolutional neural network called the Cascaded Hybrid Residual U-Net (CHR-U-Net) for MRI glioma segmentation. The CHR-U-Net exploits both the 2D local features as well as the 3D global spatial contextual information simultaneously. In the first-level of CHR-U-Net, the R-2D-U-Net combines the 2D-U-Net and the residual unit for quick lesion area detecting without any miss. To prevent from missing false-positive pixels, the output of R-2D-U-Net is resampled by using the hard-mining to collect more possible false-positive samples. In the second-level of CHR-U-Net, the axial, coronal, and sagittal 3D-U-Nets are trained to predict whether pixels belong to the area of glioma. The results of three 3D-U-Nets are fused to improve the accuracy and reduce false positives. The database of 2017 BRATS challenge were used in our experiments for the verification. The Dices and Sensitivities of Enhancing, Whole, and Core areas were calculated. The Dices are 0.73, 0.90, and 0.83 and the Sensitivities are 0.83, 0.90, and 0.82, respectively, for the axial, coronal, and sagittal 3D-U-Nets. Experimental results show that the proposed model significantly improves the performance of glioma segmentation.</description><identifier>ISSN: 1380-7501</identifier><identifier>EISSN: 1573-7721</identifier><identifier>DOI: 10.1007/s11042-020-09210-z</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Accuracy ; Artificial neural networks ; Brain cancer ; Business competition ; Computer Communication Networks ; Computer Science ; Data Structures and Information Theory ; Deep learning ; Glioma ; Image segmentation ; Machine learning ; Multimedia ; Multimedia Information Systems ; Neural networks ; Performance enhancement ; Pixels ; Sensitivity enhancement ; Special Purpose and Application-Based Systems ; Three dimensional imaging</subject><ispartof>Multimedia tools and applications, 2020-09, Vol.79 (33-34), p.24929-24947</ispartof><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020</rights><rights>Springer Science+Business Media, LLC, part of Springer Nature 2020.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c319t-935598d0c73096acb6b95b51b83d78859a5eef2bf30243d9e364b4fccc546fdb3</citedby><cites>FETCH-LOGICAL-c319t-935598d0c73096acb6b95b51b83d78859a5eef2bf30243d9e364b4fccc546fdb3</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s11042-020-09210-z$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s11042-020-09210-z$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27922,27923,41486,42555,51317</link.rule.ids></links><search><creatorcontrib>Long, Jiaosong</creatorcontrib><creatorcontrib>Ma, Guangzhi</creatorcontrib><creatorcontrib>Liu, Hong</creatorcontrib><creatorcontrib>Song, Enmin</creatorcontrib><creatorcontrib>Hung, Chih-Cheng</creatorcontrib><creatorcontrib>Xu, Xiangyang</creatorcontrib><creatorcontrib>Jin, Renchao</creatorcontrib><creatorcontrib>Zhuang, Yuzhou</creatorcontrib><creatorcontrib>Liu, DaiYang</creatorcontrib><creatorcontrib>Ma, Guangzhi</creatorcontrib><creatorcontrib>Song, Enmin</creatorcontrib><title>Cascaded hybrid residual U-Net for glioma segmentation</title><title>Multimedia tools and applications</title><addtitle>Multimed Tools Appl</addtitle><description>Glioma segmentation is critical for making surgical plans. Recently, the traditional glioma segmentation method is less competitive with two deep learning segmentation strategies: the patch-based method which focuses more on the local feature for each pixel, and the image-based method which fully leverages the global feature and captures the overall shape, size and other characteristics of the lesion in a neighborhood of a pixel. In this study, we investigate and integrate the advantages of 2-D and 3-D image-based architectures, and propose a new convolutional neural network called the Cascaded Hybrid Residual U-Net (CHR-U-Net) for MRI glioma segmentation. The CHR-U-Net exploits both the 2D local features as well as the 3D global spatial contextual information simultaneously. In the first-level of CHR-U-Net, the R-2D-U-Net combines the 2D-U-Net and the residual unit for quick lesion area detecting without any miss. To prevent from missing false-positive pixels, the output of R-2D-U-Net is resampled by using the hard-mining to collect more possible false-positive samples. In the second-level of CHR-U-Net, the axial, coronal, and sagittal 3D-U-Nets are trained to predict whether pixels belong to the area of glioma. The results of three 3D-U-Nets are fused to improve the accuracy and reduce false positives. The database of 2017 BRATS challenge were used in our experiments for the verification. The Dices and Sensitivities of Enhancing, Whole, and Core areas were calculated. The Dices are 0.73, 0.90, and 0.83 and the Sensitivities are 0.83, 0.90, and 0.82, respectively, for the axial, coronal, and sagittal 3D-U-Nets. Experimental results show that the proposed model significantly improves the performance of glioma segmentation.</description><subject>Accuracy</subject><subject>Artificial neural networks</subject><subject>Brain cancer</subject><subject>Business competition</subject><subject>Computer Communication Networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Deep learning</subject><subject>Glioma</subject><subject>Image segmentation</subject><subject>Machine learning</subject><subject>Multimedia</subject><subject>Multimedia Information Systems</subject><subject>Neural networks</subject><subject>Performance enhancement</subject><subject>Pixels</subject><subject>Sensitivity enhancement</subject><subject>Special Purpose and Application-Based Systems</subject><subject>Three dimensional imaging</subject><issn>1380-7501</issn><issn>1573-7721</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>8G5</sourceid><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GNUQQ</sourceid><sourceid>GUQSH</sourceid><sourceid>M2O</sourceid><recordid>eNp9kDtPwzAUhS0EEqXwB5giMRuu7TiOR1TxkipY6Gz5GVLlUex0aH89hiCxMd0znO9c6UPomsAtARB3iRAoKQYKGCQlgI8naEG4YFgISk5zZjVgwYGco4uUtgCk4rRcoGqlk9XOu-LjYGLriuhT6_a6Kzb41U9FGGPRdO3Y6yL5pvfDpKd2HC7RWdBd8le_d4k2jw_vq2e8fnt6Wd2vsWVETlgyzmXtwAoGstLWVEZyw4mpmRN1zaXm3gdqAgNaMic9q0pTBmstL6vgDFuim3l3F8fPvU-T2o77OOSXKgNcMsE5zS06t2wcU4o-qF1sex0PioD69qNmPyr7UT9-1DFDbIZSLg-Nj3_T_1Bf2GpnzA</recordid><startdate>20200901</startdate><enddate>20200901</enddate><creator>Long, Jiaosong</creator><creator>Ma, Guangzhi</creator><creator>Liu, Hong</creator><creator>Song, Enmin</creator><creator>Hung, Chih-Cheng</creator><creator>Xu, Xiangyang</creator><creator>Jin, Renchao</creator><creator>Zhuang, Yuzhou</creator><creator>Liu, DaiYang</creator><creator>Ma, Guangzhi</creator><creator>Song, Enmin</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8AO</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>8G5</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>GUQSH</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>M2O</scope><scope>MBDVC</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope></search><sort><creationdate>20200901</creationdate><title>Cascaded hybrid residual U-Net for glioma segmentation</title><author>Long, Jiaosong ; Ma, Guangzhi ; Liu, Hong ; Song, Enmin ; Hung, Chih-Cheng ; Xu, Xiangyang ; Jin, Renchao ; Zhuang, Yuzhou ; Liu, DaiYang ; Ma, Guangzhi ; Song, Enmin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c319t-935598d0c73096acb6b95b51b83d78859a5eef2bf30243d9e364b4fccc546fdb3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Accuracy</topic><topic>Artificial neural networks</topic><topic>Brain cancer</topic><topic>Business competition</topic><topic>Computer Communication Networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Deep learning</topic><topic>Glioma</topic><topic>Image segmentation</topic><topic>Machine learning</topic><topic>Multimedia</topic><topic>Multimedia Information Systems</topic><topic>Neural networks</topic><topic>Performance enhancement</topic><topic>Pixels</topic><topic>Sensitivity enhancement</topic><topic>Special Purpose and Application-Based Systems</topic><topic>Three dimensional imaging</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Long, Jiaosong</creatorcontrib><creatorcontrib>Ma, Guangzhi</creatorcontrib><creatorcontrib>Liu, Hong</creatorcontrib><creatorcontrib>Song, Enmin</creatorcontrib><creatorcontrib>Hung, Chih-Cheng</creatorcontrib><creatorcontrib>Xu, Xiangyang</creatorcontrib><creatorcontrib>Jin, Renchao</creatorcontrib><creatorcontrib>Zhuang, Yuzhou</creatorcontrib><creatorcontrib>Liu, DaiYang</creatorcontrib><creatorcontrib>Ma, Guangzhi</creatorcontrib><creatorcontrib>Song, Enmin</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Collection</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>ProQuest Pharma Collection</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>Research Library (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>Research Library Prep</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Research Library</collection><collection>Research Library (Corporate)</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest One Business</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Multimedia tools and applications</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Long, Jiaosong</au><au>Ma, Guangzhi</au><au>Liu, Hong</au><au>Song, Enmin</au><au>Hung, Chih-Cheng</au><au>Xu, Xiangyang</au><au>Jin, Renchao</au><au>Zhuang, Yuzhou</au><au>Liu, DaiYang</au><au>Ma, Guangzhi</au><au>Song, Enmin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Cascaded hybrid residual U-Net for glioma segmentation</atitle><jtitle>Multimedia tools and applications</jtitle><stitle>Multimed Tools Appl</stitle><date>2020-09-01</date><risdate>2020</risdate><volume>79</volume><issue>33-34</issue><spage>24929</spage><epage>24947</epage><pages>24929-24947</pages><issn>1380-7501</issn><eissn>1573-7721</eissn><abstract>Glioma segmentation is critical for making surgical plans. Recently, the traditional glioma segmentation method is less competitive with two deep learning segmentation strategies: the patch-based method which focuses more on the local feature for each pixel, and the image-based method which fully leverages the global feature and captures the overall shape, size and other characteristics of the lesion in a neighborhood of a pixel. In this study, we investigate and integrate the advantages of 2-D and 3-D image-based architectures, and propose a new convolutional neural network called the Cascaded Hybrid Residual U-Net (CHR-U-Net) for MRI glioma segmentation. The CHR-U-Net exploits both the 2D local features as well as the 3D global spatial contextual information simultaneously. In the first-level of CHR-U-Net, the R-2D-U-Net combines the 2D-U-Net and the residual unit for quick lesion area detecting without any miss. To prevent from missing false-positive pixels, the output of R-2D-U-Net is resampled by using the hard-mining to collect more possible false-positive samples. In the second-level of CHR-U-Net, the axial, coronal, and sagittal 3D-U-Nets are trained to predict whether pixels belong to the area of glioma. The results of three 3D-U-Nets are fused to improve the accuracy and reduce false positives. The database of 2017 BRATS challenge were used in our experiments for the verification. The Dices and Sensitivities of Enhancing, Whole, and Core areas were calculated. The Dices are 0.73, 0.90, and 0.83 and the Sensitivities are 0.83, 0.90, and 0.82, respectively, for the axial, coronal, and sagittal 3D-U-Nets. Experimental results show that the proposed model significantly improves the performance of glioma segmentation.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s11042-020-09210-z</doi><tpages>19</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1380-7501
ispartof Multimedia tools and applications, 2020-09, Vol.79 (33-34), p.24929-24947
issn 1380-7501
1573-7721
language eng
recordid cdi_proquest_journals_2435937552
source SpringerLink Journals - AutoHoldings
subjects Accuracy
Artificial neural networks
Brain cancer
Business competition
Computer Communication Networks
Computer Science
Data Structures and Information Theory
Deep learning
Glioma
Image segmentation
Machine learning
Multimedia
Multimedia Information Systems
Neural networks
Performance enhancement
Pixels
Sensitivity enhancement
Special Purpose and Application-Based Systems
Three dimensional imaging
title Cascaded hybrid residual U-Net for glioma segmentation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-09T21%3A43%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Cascaded%20hybrid%20residual%20U-Net%20for%20glioma%20segmentation&rft.jtitle=Multimedia%20tools%20and%20applications&rft.au=Long,%20Jiaosong&rft.date=2020-09-01&rft.volume=79&rft.issue=33-34&rft.spage=24929&rft.epage=24947&rft.pages=24929-24947&rft.issn=1380-7501&rft.eissn=1573-7721&rft_id=info:doi/10.1007/s11042-020-09210-z&rft_dat=%3Cproquest_cross%3E2435937552%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2435937552&rft_id=info:pmid/&rfr_iscdi=true