Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks

Purpose Dual‐energy computed tomography (DECT) has shown great potential in many clinical applications. By incorporating the information from two different energy spectra, DECT provides higher contrast and reveals more material differences of tissues compared to conventional single‐energy CT (SECT)....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Medical physics (Lancaster) 2020-02, Vol.47 (2), p.552-562
Hauptverfasser: Chen, Shuqing, Zhong, Xia, Hu, Shiyang, Dorn, Sabrina, Kachelrieß, Marc, Lell, Michael, Maier, Andreas
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 562
container_issue 2
container_start_page 552
container_title Medical physics (Lancaster)
container_volume 47
creator Chen, Shuqing
Zhong, Xia
Hu, Shiyang
Dorn, Sabrina
Kachelrieß, Marc
Lell, Michael
Maier, Andreas
description Purpose Dual‐energy computed tomography (DECT) has shown great potential in many clinical applications. By incorporating the information from two different energy spectra, DECT provides higher contrast and reveals more material differences of tissues compared to conventional single‐energy CT (SECT). Recent research shows that automatic multi‐organ segmentation of DECT data can improve DECT clinical applications. However, most segmentation methods are designed for SECT, while DECT has been significantly less pronounced in research. Therefore, a novel approach is required that is able to take full advantage of the extra information provided by DECT. Methods In the scope of this work, we proposed four three‐dimensional (3D) fully convolutional neural network algorithms for the automatic segmentation of DECT data. We incorporated the extra energy information differently and embedded the fusion of information in each of the network architectures. Results Quantitative evaluation using 45 thorax/abdomen DECT datasets acquired with a clinical dual‐source CT system was investigated. The segmentation of six thoracic and abdominal organs (left and right lungs, liver, spleen, and left and right kidneys) were evaluated using a fivefold cross‐validation strategy. In all of the tests, we achieved the best average Dice coefficients of 98% for the right lung, 98% for the left lung, 96% for the liver, 92% for the spleen, 95% for the right kidney, 93% for the left kidney, respectively. The network architectures exploit dual‐energy spectra and outperform deep learning for SECT. Conclusions The results of the cross‐validation show that our methods are feasible and promising. Successful tests on special clinical cases reveal that our methods have high adaptability in the practical application.
doi_str_mv 10.1002/mp.13950
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2323481636</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2323481636</sourcerecordid><originalsourceid>FETCH-LOGICAL-c3550-5abec2ac48fd6e572d9c165d3069f30271402876398b81dbd33b4b4a3257b3633</originalsourceid><addsrcrecordid>eNp1kLtOwzAUhi0EgnKReALkEYbAsY-dy4habhIIhjJHju2UgBOXOKHqxiPwjDwJKQU2pjP83_mGj5BDBqcMgJ_V81OGmYQNMuIiwUhwyDbJCCATERcgd8huCM8AEKOEbbKDLGUxZHJEwnnf-Vp1laZ177rq8_3DtzPV0GBntW26YfENrRpqeuWG0Ta2nS3peEqPJxfj6QldVN0TNdZUWnXWUJzQsnduSbVv3rzrV-_K0RVLG9stfPsS9slWqVywBz93jzxeXkzH19Ht_dXN-Pw20iglRFIVVnOlRVqa2MqEm0yzWBqEOCsReMIE8DSJMUuLlJnCIBaiEAq5TAqMEffI8do7b_1rb0OX11XQ1jnVWN-HnCNHMYQY2D9Utz6E1pb5vK1q1S5zBvkqcV7P8-_EA3r0Y-2L2po_8LfpAERrYFE5u_xXlN89rIVflZCFkw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2323481636</pqid></control><display><type>article</type><title>Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks</title><source>Wiley Online Library All Journals</source><source>Alma/SFX Local Collection</source><creator>Chen, Shuqing ; Zhong, Xia ; Hu, Shiyang ; Dorn, Sabrina ; Kachelrieß, Marc ; Lell, Michael ; Maier, Andreas</creator><creatorcontrib>Chen, Shuqing ; Zhong, Xia ; Hu, Shiyang ; Dorn, Sabrina ; Kachelrieß, Marc ; Lell, Michael ; Maier, Andreas</creatorcontrib><description>Purpose Dual‐energy computed tomography (DECT) has shown great potential in many clinical applications. By incorporating the information from two different energy spectra, DECT provides higher contrast and reveals more material differences of tissues compared to conventional single‐energy CT (SECT). Recent research shows that automatic multi‐organ segmentation of DECT data can improve DECT clinical applications. However, most segmentation methods are designed for SECT, while DECT has been significantly less pronounced in research. Therefore, a novel approach is required that is able to take full advantage of the extra information provided by DECT. Methods In the scope of this work, we proposed four three‐dimensional (3D) fully convolutional neural network algorithms for the automatic segmentation of DECT data. We incorporated the extra energy information differently and embedded the fusion of information in each of the network architectures. Results Quantitative evaluation using 45 thorax/abdomen DECT datasets acquired with a clinical dual‐source CT system was investigated. The segmentation of six thoracic and abdominal organs (left and right lungs, liver, spleen, and left and right kidneys) were evaluated using a fivefold cross‐validation strategy. In all of the tests, we achieved the best average Dice coefficients of 98% for the right lung, 98% for the left lung, 96% for the liver, 92% for the spleen, 95% for the right kidney, 93% for the left kidney, respectively. The network architectures exploit dual‐energy spectra and outperform deep learning for SECT. Conclusions The results of the cross‐validation show that our methods are feasible and promising. Successful tests on special clinical cases reveal that our methods have high adaptability in the practical application.</description><identifier>ISSN: 0094-2405</identifier><identifier>EISSN: 2473-4209</identifier><identifier>DOI: 10.1002/mp.13950</identifier><identifier>PMID: 31816095</identifier><language>eng</language><publisher>United States</publisher><subject>DECT ; deep learning ; FCN ; multi‐organ ; U‐Net</subject><ispartof>Medical physics (Lancaster), 2020-02, Vol.47 (2), p.552-562</ispartof><rights>2019 The Authors. Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c3550-5abec2ac48fd6e572d9c165d3069f30271402876398b81dbd33b4b4a3257b3633</citedby><cites>FETCH-LOGICAL-c3550-5abec2ac48fd6e572d9c165d3069f30271402876398b81dbd33b4b4a3257b3633</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://onlinelibrary.wiley.com/doi/pdf/10.1002%2Fmp.13950$$EPDF$$P50$$Gwiley$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://onlinelibrary.wiley.com/doi/full/10.1002%2Fmp.13950$$EHTML$$P50$$Gwiley$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,1417,27924,27925,45574,45575</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/31816095$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Shuqing</creatorcontrib><creatorcontrib>Zhong, Xia</creatorcontrib><creatorcontrib>Hu, Shiyang</creatorcontrib><creatorcontrib>Dorn, Sabrina</creatorcontrib><creatorcontrib>Kachelrieß, Marc</creatorcontrib><creatorcontrib>Lell, Michael</creatorcontrib><creatorcontrib>Maier, Andreas</creatorcontrib><title>Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks</title><title>Medical physics (Lancaster)</title><addtitle>Med Phys</addtitle><description>Purpose Dual‐energy computed tomography (DECT) has shown great potential in many clinical applications. By incorporating the information from two different energy spectra, DECT provides higher contrast and reveals more material differences of tissues compared to conventional single‐energy CT (SECT). Recent research shows that automatic multi‐organ segmentation of DECT data can improve DECT clinical applications. However, most segmentation methods are designed for SECT, while DECT has been significantly less pronounced in research. Therefore, a novel approach is required that is able to take full advantage of the extra information provided by DECT. Methods In the scope of this work, we proposed four three‐dimensional (3D) fully convolutional neural network algorithms for the automatic segmentation of DECT data. We incorporated the extra energy information differently and embedded the fusion of information in each of the network architectures. Results Quantitative evaluation using 45 thorax/abdomen DECT datasets acquired with a clinical dual‐source CT system was investigated. The segmentation of six thoracic and abdominal organs (left and right lungs, liver, spleen, and left and right kidneys) were evaluated using a fivefold cross‐validation strategy. In all of the tests, we achieved the best average Dice coefficients of 98% for the right lung, 98% for the left lung, 96% for the liver, 92% for the spleen, 95% for the right kidney, 93% for the left kidney, respectively. The network architectures exploit dual‐energy spectra and outperform deep learning for SECT. Conclusions The results of the cross‐validation show that our methods are feasible and promising. Successful tests on special clinical cases reveal that our methods have high adaptability in the practical application.</description><subject>DECT</subject><subject>deep learning</subject><subject>FCN</subject><subject>multi‐organ</subject><subject>U‐Net</subject><issn>0094-2405</issn><issn>2473-4209</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>24P</sourceid><sourceid>WIN</sourceid><recordid>eNp1kLtOwzAUhi0EgnKReALkEYbAsY-dy4habhIIhjJHju2UgBOXOKHqxiPwjDwJKQU2pjP83_mGj5BDBqcMgJ_V81OGmYQNMuIiwUhwyDbJCCATERcgd8huCM8AEKOEbbKDLGUxZHJEwnnf-Vp1laZ177rq8_3DtzPV0GBntW26YfENrRpqeuWG0Ta2nS3peEqPJxfj6QldVN0TNdZUWnXWUJzQsnduSbVv3rzrV-_K0RVLG9stfPsS9slWqVywBz93jzxeXkzH19Ht_dXN-Pw20iglRFIVVnOlRVqa2MqEm0yzWBqEOCsReMIE8DSJMUuLlJnCIBaiEAq5TAqMEffI8do7b_1rb0OX11XQ1jnVWN-HnCNHMYQY2D9Utz6E1pb5vK1q1S5zBvkqcV7P8-_EA3r0Y-2L2po_8LfpAERrYFE5u_xXlN89rIVflZCFkw</recordid><startdate>202002</startdate><enddate>202002</enddate><creator>Chen, Shuqing</creator><creator>Zhong, Xia</creator><creator>Hu, Shiyang</creator><creator>Dorn, Sabrina</creator><creator>Kachelrieß, Marc</creator><creator>Lell, Michael</creator><creator>Maier, Andreas</creator><scope>24P</scope><scope>WIN</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope></search><sort><creationdate>202002</creationdate><title>Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks</title><author>Chen, Shuqing ; Zhong, Xia ; Hu, Shiyang ; Dorn, Sabrina ; Kachelrieß, Marc ; Lell, Michael ; Maier, Andreas</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c3550-5abec2ac48fd6e572d9c165d3069f30271402876398b81dbd33b4b4a3257b3633</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>DECT</topic><topic>deep learning</topic><topic>FCN</topic><topic>multi‐organ</topic><topic>U‐Net</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chen, Shuqing</creatorcontrib><creatorcontrib>Zhong, Xia</creatorcontrib><creatorcontrib>Hu, Shiyang</creatorcontrib><creatorcontrib>Dorn, Sabrina</creatorcontrib><creatorcontrib>Kachelrieß, Marc</creatorcontrib><creatorcontrib>Lell, Michael</creatorcontrib><creatorcontrib>Maier, Andreas</creatorcontrib><collection>Wiley-Blackwell Open Access Titles</collection><collection>Wiley Free Content</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Medical physics (Lancaster)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chen, Shuqing</au><au>Zhong, Xia</au><au>Hu, Shiyang</au><au>Dorn, Sabrina</au><au>Kachelrieß, Marc</au><au>Lell, Michael</au><au>Maier, Andreas</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks</atitle><jtitle>Medical physics (Lancaster)</jtitle><addtitle>Med Phys</addtitle><date>2020-02</date><risdate>2020</risdate><volume>47</volume><issue>2</issue><spage>552</spage><epage>562</epage><pages>552-562</pages><issn>0094-2405</issn><eissn>2473-4209</eissn><abstract>Purpose Dual‐energy computed tomography (DECT) has shown great potential in many clinical applications. By incorporating the information from two different energy spectra, DECT provides higher contrast and reveals more material differences of tissues compared to conventional single‐energy CT (SECT). Recent research shows that automatic multi‐organ segmentation of DECT data can improve DECT clinical applications. However, most segmentation methods are designed for SECT, while DECT has been significantly less pronounced in research. Therefore, a novel approach is required that is able to take full advantage of the extra information provided by DECT. Methods In the scope of this work, we proposed four three‐dimensional (3D) fully convolutional neural network algorithms for the automatic segmentation of DECT data. We incorporated the extra energy information differently and embedded the fusion of information in each of the network architectures. Results Quantitative evaluation using 45 thorax/abdomen DECT datasets acquired with a clinical dual‐source CT system was investigated. The segmentation of six thoracic and abdominal organs (left and right lungs, liver, spleen, and left and right kidneys) were evaluated using a fivefold cross‐validation strategy. In all of the tests, we achieved the best average Dice coefficients of 98% for the right lung, 98% for the left lung, 96% for the liver, 92% for the spleen, 95% for the right kidney, 93% for the left kidney, respectively. The network architectures exploit dual‐energy spectra and outperform deep learning for SECT. Conclusions The results of the cross‐validation show that our methods are feasible and promising. Successful tests on special clinical cases reveal that our methods have high adaptability in the practical application.</abstract><cop>United States</cop><pmid>31816095</pmid><doi>10.1002/mp.13950</doi><tpages>11</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 0094-2405
ispartof Medical physics (Lancaster), 2020-02, Vol.47 (2), p.552-562
issn 0094-2405
2473-4209
language eng
recordid cdi_proquest_miscellaneous_2323481636
source Wiley Online Library All Journals; Alma/SFX Local Collection
subjects DECT
deep learning
FCN
multi‐organ
U‐Net
title Automatic multi‐organ segmentation in dual‐energy CT (DECT) with dedicated 3D fully convolutional DECT networks
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T04%3A04%3A56IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Automatic%20multi%E2%80%90organ%20segmentation%20in%20dual%E2%80%90energy%20CT%20(DECT)%20with%20dedicated%203D%20fully%20convolutional%20DECT%20networks&rft.jtitle=Medical%20physics%20(Lancaster)&rft.au=Chen,%20Shuqing&rft.date=2020-02&rft.volume=47&rft.issue=2&rft.spage=552&rft.epage=562&rft.pages=552-562&rft.issn=0094-2405&rft.eissn=2473-4209&rft_id=info:doi/10.1002/mp.13950&rft_dat=%3Cproquest_cross%3E2323481636%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2323481636&rft_id=info:pmid/31816095&rfr_iscdi=true