Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging

Purpose To develop a deep neural network system for the automatic segmentation and risk stratification prediction of gastrointestinal stromal tumors (GISTs). Methods A total of 980 ultrasound (US) images from 245 GIST patients were retrospectively collected. These images were randomly divided (6:2:2...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of ultrasound in medicine 2024-09, Vol.43 (9), p.1661-1672
Hauptverfasser: Zhuo, Minling, Chen, Xing, Guo, Jingjing, Qian, Qingfu, Xue, Ensheng, Chen, Zhikui
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1672
container_issue 9
container_start_page 1661
container_title Journal of ultrasound in medicine
container_volume 43
creator Zhuo, Minling
Chen, Xing
Guo, Jingjing
Qian, Qingfu
Xue, Ensheng
Chen, Zhikui
description Purpose To develop a deep neural network system for the automatic segmentation and risk stratification prediction of gastrointestinal stromal tumors (GISTs). Methods A total of 980 ultrasound (US) images from 245 GIST patients were retrospectively collected. These images were randomly divided (6:2:2) into a training set, a validation set, and an internal test set. Additionally, 188 US images from 47 prospective GIST patients were collected to evaluate the segmentation and diagnostic performance of the model. Five deep learning‐based segmentation networks, namely, UNet, FCN, DeepLabV3+, Swin Transformer, and SegNeXt, were employed, along with the ResNet 18 classification network, to select the most suitable network combination. The performance of the segmentation models was evaluated using metrics such as the intersection over union (IoU), Dice similarity coefficient (DSC), recall, and precision. The classification performance was assessed based on accuracy and the area under the receiver operating characteristic curve (AUROC). Results Among the compared models, SegNeXt‐ResNet18 exhibited the best segmentation and classification performance. On the internal test set, the proposed model achieved IoU, DSC, precision, and recall values of 82.1, 90.2, 91.7, and 88.8%, respectively. The accuracy and AUC for GIST risk prediction were 87.4 and 92.0%, respectively. On the external test set, the segmentation models exhibited IoU, DSC, precision, and recall values of 81.0, 89.5, 92.8, and 86.4%, respectively. The accuracy and AUC for GIST risk prediction were 86.7 and 92.5%, respectively. Conclusion This two‐stage SegNeXt‐ResNet18 model achieves automatic segmentation and risk stratification prediction for GISTs and demonstrates excellent segmentation and classification performance.
doi_str_mv 10.1002/jum.16489
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_3063465747</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3063465747</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2159-3300c47d6fa532dc1d5cc2f035abc814deaffd3ef85ea6fb87689e955b3a3f623</originalsourceid><addsrcrecordid>eNp1kM1u1DAUhS0EokNhwQugLGGR1j9xfpZQSikahERn1tGNfT1yie3BTlR1V6kvwDPyJLhN6a6rIx19-u7VIeQto0eMUn58ObsjVldt94ysmJS07GomnpMV5U1bVrxrDsirlC4zSllTvSQHom05Z51ckdvPiPtijRC99bu_N38-QUJdXODOoZ9gssEX4HXx06ZfxcUUc2OsWnoTYnEGaYrB-gnTZD2Md0xwOTezCzEV1hebCD7BoIO7B7ZjtqQwZ-m5g12--pq8MDAmfPOQh2T75XRz8rVc_zg7P_m4LhVnsiuFoFRVja4NSMG1YloqxQ0VEgbVskojGKMFmlYi1GZom7rtsJNyECBMzcUheb949zH8nvPDvbNJ4TiCxzCnXtBaVLVsqiajHxZUxZBSRNPvo3UQr3tG-7vN-7x5f795Zt89aOfBoX4k_4-cgeMFuLIjXj9t6r9tvy_Kf1l1j9k</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3063465747</pqid></control><display><type>article</type><title>Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging</title><source>Wiley Journals</source><creator>Zhuo, Minling ; Chen, Xing ; Guo, Jingjing ; Qian, Qingfu ; Xue, Ensheng ; Chen, Zhikui</creator><creatorcontrib>Zhuo, Minling ; Chen, Xing ; Guo, Jingjing ; Qian, Qingfu ; Xue, Ensheng ; Chen, Zhikui</creatorcontrib><description>Purpose To develop a deep neural network system for the automatic segmentation and risk stratification prediction of gastrointestinal stromal tumors (GISTs). Methods A total of 980 ultrasound (US) images from 245 GIST patients were retrospectively collected. These images were randomly divided (6:2:2) into a training set, a validation set, and an internal test set. Additionally, 188 US images from 47 prospective GIST patients were collected to evaluate the segmentation and diagnostic performance of the model. Five deep learning‐based segmentation networks, namely, UNet, FCN, DeepLabV3+, Swin Transformer, and SegNeXt, were employed, along with the ResNet 18 classification network, to select the most suitable network combination. The performance of the segmentation models was evaluated using metrics such as the intersection over union (IoU), Dice similarity coefficient (DSC), recall, and precision. The classification performance was assessed based on accuracy and the area under the receiver operating characteristic curve (AUROC). Results Among the compared models, SegNeXt‐ResNet18 exhibited the best segmentation and classification performance. On the internal test set, the proposed model achieved IoU, DSC, precision, and recall values of 82.1, 90.2, 91.7, and 88.8%, respectively. The accuracy and AUC for GIST risk prediction were 87.4 and 92.0%, respectively. On the external test set, the segmentation models exhibited IoU, DSC, precision, and recall values of 81.0, 89.5, 92.8, and 86.4%, respectively. The accuracy and AUC for GIST risk prediction were 86.7 and 92.5%, respectively. Conclusion This two‐stage SegNeXt‐ResNet18 model achieves automatic segmentation and risk stratification prediction for GISTs and demonstrates excellent segmentation and classification performance.</description><identifier>ISSN: 0278-4297</identifier><identifier>ISSN: 1550-9613</identifier><identifier>EISSN: 1550-9613</identifier><identifier>DOI: 10.1002/jum.16489</identifier><identifier>PMID: 38822195</identifier><language>eng</language><publisher>Hoboken, USA: John Wiley &amp; Sons, Inc</publisher><subject>deep learning ; gastrointestinal stromal tumor ; risk classification ; segmentation ; ultrasound image</subject><ispartof>Journal of ultrasound in medicine, 2024-09, Vol.43 (9), p.1661-1672</ispartof><rights>2024 American Institute of Ultrasound in Medicine.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2159-3300c47d6fa532dc1d5cc2f035abc814deaffd3ef85ea6fb87689e955b3a3f623</cites><orcidid>0000-0002-8998-5437 ; 0000-0001-6941-9641</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://onlinelibrary.wiley.com/doi/pdf/10.1002%2Fjum.16489$$EPDF$$P50$$Gwiley$$H</linktopdf><linktohtml>$$Uhttps://onlinelibrary.wiley.com/doi/full/10.1002%2Fjum.16489$$EHTML$$P50$$Gwiley$$H</linktohtml><link.rule.ids>314,780,784,1417,27924,27925,45574,45575</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/38822195$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Zhuo, Minling</creatorcontrib><creatorcontrib>Chen, Xing</creatorcontrib><creatorcontrib>Guo, Jingjing</creatorcontrib><creatorcontrib>Qian, Qingfu</creatorcontrib><creatorcontrib>Xue, Ensheng</creatorcontrib><creatorcontrib>Chen, Zhikui</creatorcontrib><title>Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging</title><title>Journal of ultrasound in medicine</title><addtitle>J Ultrasound Med</addtitle><description>Purpose To develop a deep neural network system for the automatic segmentation and risk stratification prediction of gastrointestinal stromal tumors (GISTs). Methods A total of 980 ultrasound (US) images from 245 GIST patients were retrospectively collected. These images were randomly divided (6:2:2) into a training set, a validation set, and an internal test set. Additionally, 188 US images from 47 prospective GIST patients were collected to evaluate the segmentation and diagnostic performance of the model. Five deep learning‐based segmentation networks, namely, UNet, FCN, DeepLabV3+, Swin Transformer, and SegNeXt, were employed, along with the ResNet 18 classification network, to select the most suitable network combination. The performance of the segmentation models was evaluated using metrics such as the intersection over union (IoU), Dice similarity coefficient (DSC), recall, and precision. The classification performance was assessed based on accuracy and the area under the receiver operating characteristic curve (AUROC). Results Among the compared models, SegNeXt‐ResNet18 exhibited the best segmentation and classification performance. On the internal test set, the proposed model achieved IoU, DSC, precision, and recall values of 82.1, 90.2, 91.7, and 88.8%, respectively. The accuracy and AUC for GIST risk prediction were 87.4 and 92.0%, respectively. On the external test set, the segmentation models exhibited IoU, DSC, precision, and recall values of 81.0, 89.5, 92.8, and 86.4%, respectively. The accuracy and AUC for GIST risk prediction were 86.7 and 92.5%, respectively. Conclusion This two‐stage SegNeXt‐ResNet18 model achieves automatic segmentation and risk stratification prediction for GISTs and demonstrates excellent segmentation and classification performance.</description><subject>deep learning</subject><subject>gastrointestinal stromal tumor</subject><subject>risk classification</subject><subject>segmentation</subject><subject>ultrasound image</subject><issn>0278-4297</issn><issn>1550-9613</issn><issn>1550-9613</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp1kM1u1DAUhS0EokNhwQugLGGR1j9xfpZQSikahERn1tGNfT1yie3BTlR1V6kvwDPyJLhN6a6rIx19-u7VIeQto0eMUn58ObsjVldt94ysmJS07GomnpMV5U1bVrxrDsirlC4zSllTvSQHom05Z51ckdvPiPtijRC99bu_N38-QUJdXODOoZ9gssEX4HXx06ZfxcUUc2OsWnoTYnEGaYrB-gnTZD2Md0xwOTezCzEV1hebCD7BoIO7B7ZjtqQwZ-m5g12--pq8MDAmfPOQh2T75XRz8rVc_zg7P_m4LhVnsiuFoFRVja4NSMG1YloqxQ0VEgbVskojGKMFmlYi1GZom7rtsJNyECBMzcUheb949zH8nvPDvbNJ4TiCxzCnXtBaVLVsqiajHxZUxZBSRNPvo3UQr3tG-7vN-7x5f795Zt89aOfBoX4k_4-cgeMFuLIjXj9t6r9tvy_Kf1l1j9k</recordid><startdate>202409</startdate><enddate>202409</enddate><creator>Zhuo, Minling</creator><creator>Chen, Xing</creator><creator>Guo, Jingjing</creator><creator>Qian, Qingfu</creator><creator>Xue, Ensheng</creator><creator>Chen, Zhikui</creator><general>John Wiley &amp; Sons, Inc</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0002-8998-5437</orcidid><orcidid>https://orcid.org/0000-0001-6941-9641</orcidid></search><sort><creationdate>202409</creationdate><title>Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging</title><author>Zhuo, Minling ; Chen, Xing ; Guo, Jingjing ; Qian, Qingfu ; Xue, Ensheng ; Chen, Zhikui</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2159-3300c47d6fa532dc1d5cc2f035abc814deaffd3ef85ea6fb87689e955b3a3f623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>deep learning</topic><topic>gastrointestinal stromal tumor</topic><topic>risk classification</topic><topic>segmentation</topic><topic>ultrasound image</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhuo, Minling</creatorcontrib><creatorcontrib>Chen, Xing</creatorcontrib><creatorcontrib>Guo, Jingjing</creatorcontrib><creatorcontrib>Qian, Qingfu</creatorcontrib><creatorcontrib>Xue, Ensheng</creatorcontrib><creatorcontrib>Chen, Zhikui</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Journal of ultrasound in medicine</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhuo, Minling</au><au>Chen, Xing</au><au>Guo, Jingjing</au><au>Qian, Qingfu</au><au>Xue, Ensheng</au><au>Chen, Zhikui</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging</atitle><jtitle>Journal of ultrasound in medicine</jtitle><addtitle>J Ultrasound Med</addtitle><date>2024-09</date><risdate>2024</risdate><volume>43</volume><issue>9</issue><spage>1661</spage><epage>1672</epage><pages>1661-1672</pages><issn>0278-4297</issn><issn>1550-9613</issn><eissn>1550-9613</eissn><abstract>Purpose To develop a deep neural network system for the automatic segmentation and risk stratification prediction of gastrointestinal stromal tumors (GISTs). Methods A total of 980 ultrasound (US) images from 245 GIST patients were retrospectively collected. These images were randomly divided (6:2:2) into a training set, a validation set, and an internal test set. Additionally, 188 US images from 47 prospective GIST patients were collected to evaluate the segmentation and diagnostic performance of the model. Five deep learning‐based segmentation networks, namely, UNet, FCN, DeepLabV3+, Swin Transformer, and SegNeXt, were employed, along with the ResNet 18 classification network, to select the most suitable network combination. The performance of the segmentation models was evaluated using metrics such as the intersection over union (IoU), Dice similarity coefficient (DSC), recall, and precision. The classification performance was assessed based on accuracy and the area under the receiver operating characteristic curve (AUROC). Results Among the compared models, SegNeXt‐ResNet18 exhibited the best segmentation and classification performance. On the internal test set, the proposed model achieved IoU, DSC, precision, and recall values of 82.1, 90.2, 91.7, and 88.8%, respectively. The accuracy and AUC for GIST risk prediction were 87.4 and 92.0%, respectively. On the external test set, the segmentation models exhibited IoU, DSC, precision, and recall values of 81.0, 89.5, 92.8, and 86.4%, respectively. The accuracy and AUC for GIST risk prediction were 86.7 and 92.5%, respectively. Conclusion This two‐stage SegNeXt‐ResNet18 model achieves automatic segmentation and risk stratification prediction for GISTs and demonstrates excellent segmentation and classification performance.</abstract><cop>Hoboken, USA</cop><pub>John Wiley &amp; Sons, Inc</pub><pmid>38822195</pmid><doi>10.1002/jum.16489</doi><tpages>12</tpages><orcidid>https://orcid.org/0000-0002-8998-5437</orcidid><orcidid>https://orcid.org/0000-0001-6941-9641</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0278-4297
ispartof Journal of ultrasound in medicine, 2024-09, Vol.43 (9), p.1661-1672
issn 0278-4297
1550-9613
1550-9613
language eng
recordid cdi_proquest_miscellaneous_3063465747
source Wiley Journals
subjects deep learning
gastrointestinal stromal tumor
risk classification
segmentation
ultrasound image
title Deep Learning‐Based Segmentation and Risk Stratification for Gastrointestinal Stromal Tumors in Transabdominal Ultrasound Imaging
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T04%3A49%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20Learning%E2%80%90Based%20Segmentation%20and%20Risk%20Stratification%20for%20Gastrointestinal%20Stromal%20Tumors%20in%20Transabdominal%20Ultrasound%20Imaging&rft.jtitle=Journal%20of%20ultrasound%20in%20medicine&rft.au=Zhuo,%20Minling&rft.date=2024-09&rft.volume=43&rft.issue=9&rft.spage=1661&rft.epage=1672&rft.pages=1661-1672&rft.issn=0278-4297&rft.eissn=1550-9613&rft_id=info:doi/10.1002/jum.16489&rft_dat=%3Cproquest_cross%3E3063465747%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3063465747&rft_id=info:pmid/38822195&rfr_iscdi=true