Diverse data augmentation for learning image segmentation with cross-modality annotations

•We propose a novel cross-modality medical image segmentation method.•Our model can perform segmentation for a target domain without labeled training data.•A diverse data augmentation approach is used to augment the training data for segmentation.•Experiments in two different tasks demonstrate the e...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Medical image analysis 2021-07, Vol.71, p.102060-102060, Article 102060
Hauptverfasser: Chen, Xu, Lian, Chunfeng, Wang, Li, Deng, Hannah, Kuang, Tianshu, Fung, Steve H., Gateno, Jaime, Shen, Dinggang, Xia, James J., Yap, Pew-Thian
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 102060
container_issue
container_start_page 102060
container_title Medical image analysis
container_volume 71
creator Chen, Xu
Lian, Chunfeng
Wang, Li
Deng, Hannah
Kuang, Tianshu
Fung, Steve H.
Gateno, Jaime
Shen, Dinggang
Xia, James J.
Yap, Pew-Thian
description •We propose a novel cross-modality medical image segmentation method.•Our model can perform segmentation for a target domain without labeled training data.•A diverse data augmentation approach is used to augment the training data for segmentation.•Experiments in two different tasks demonstrate the effectiveness of proposed method. [Display omitted] The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.
doi_str_mv 10.1016/j.media.2021.102060
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_8184609</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><els_id>S1361841521001067</els_id><sourcerecordid>2524351164</sourcerecordid><originalsourceid>FETCH-LOGICAL-c487t-d95f3e10407dd42e9e271c65b5da0d47b5fbde1e6bee12195460597849a842ad3</originalsourceid><addsrcrecordid>eNp9kU9P3DAQxa2qqGyBT1CpitQLlyz-n-QAUgUtIK3EBQ6cLCeeLF4lNtjJVnx7vBu6Ag6cbPn95nlmHkI_CJ4TTOTJat6DsXpOMSXphWKJv6AZYZLkJafs6-5OxD76HuMKY1xwjr-hfcYqUQhRztD9hV1DiJAZPehMj8se3KAH613W-pB1oIOzbpnZXi8hi_BG_2eHh6wJPsa890Z3dnjOtHN-kuMh2mt1F-Ho9TxAd3__3J5f5Yuby-vz34u84WUx5KYSLQOCOS6M4RQqoAVppKiF0djwohZtbYCArAEIJZXgEouqKHml05TasAN0Nvk-jnVaSJP6C7pTjyG1HJ6V11a9V5x9UEu_ViUpk1eVDI5fDYJ_GiEOqrexga7TDvwYFRWUM0GI5An99QFd-TG4NF6iJOOFTGCi2ERtlxOg3TVDsNpEp1ZqG53aRKem6FLVz7dz7Gr-Z5WA0wmAtM21haBiY8E1ySlAMyjj7acfvAAfvKzO</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2563476351</pqid></control><display><type>article</type><title>Diverse data augmentation for learning image segmentation with cross-modality annotations</title><source>Elsevier ScienceDirect Journals Complete</source><creator>Chen, Xu ; Lian, Chunfeng ; Wang, Li ; Deng, Hannah ; Kuang, Tianshu ; Fung, Steve H. ; Gateno, Jaime ; Shen, Dinggang ; Xia, James J. ; Yap, Pew-Thian</creator><creatorcontrib>Chen, Xu ; Lian, Chunfeng ; Wang, Li ; Deng, Hannah ; Kuang, Tianshu ; Fung, Steve H. ; Gateno, Jaime ; Shen, Dinggang ; Xia, James J. ; Yap, Pew-Thian</creatorcontrib><description>•We propose a novel cross-modality medical image segmentation method.•Our model can perform segmentation for a target domain without labeled training data.•A diverse data augmentation approach is used to augment the training data for segmentation.•Experiments in two different tasks demonstrate the effectiveness of proposed method. [Display omitted] The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.</description><identifier>ISSN: 1361-8415</identifier><identifier>EISSN: 1361-8423</identifier><identifier>DOI: 10.1016/j.media.2021.102060</identifier><identifier>PMID: 33957558</identifier><language>eng</language><publisher>Netherlands: Elsevier B.V</publisher><subject>Annotations ; Artificial neural networks ; Computed tomography ; Data augmentation ; Disentangled representation learning ; Generative adversarial learning ; Generative adversarial networks ; Image processing ; Image segmentation ; Magnetic resonance imaging ; Medical image segmentation ; Medical imaging ; Neural networks</subject><ispartof>Medical image analysis, 2021-07, Vol.71, p.102060-102060, Article 102060</ispartof><rights>2021 Elsevier B.V.</rights><rights>Copyright © 2021 Elsevier B.V. All rights reserved.</rights><rights>Copyright Elsevier BV Jul 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c487t-d95f3e10407dd42e9e271c65b5da0d47b5fbde1e6bee12195460597849a842ad3</citedby><cites>FETCH-LOGICAL-c487t-d95f3e10407dd42e9e271c65b5da0d47b5fbde1e6bee12195460597849a842ad3</cites><orcidid>0000-0001-9363-1722 ; 0000-0002-0367-3003 ; 0000-0002-1177-682X ; 0000-0002-9319-6633 ; 0000-0003-1489-2102</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://www.sciencedirect.com/science/article/pii/S1361841521001067$$EHTML$$P50$$Gelsevier$$H</linktohtml><link.rule.ids>230,314,776,780,881,3537,27901,27902,65534</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/33957558$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Xu</creatorcontrib><creatorcontrib>Lian, Chunfeng</creatorcontrib><creatorcontrib>Wang, Li</creatorcontrib><creatorcontrib>Deng, Hannah</creatorcontrib><creatorcontrib>Kuang, Tianshu</creatorcontrib><creatorcontrib>Fung, Steve H.</creatorcontrib><creatorcontrib>Gateno, Jaime</creatorcontrib><creatorcontrib>Shen, Dinggang</creatorcontrib><creatorcontrib>Xia, James J.</creatorcontrib><creatorcontrib>Yap, Pew-Thian</creatorcontrib><title>Diverse data augmentation for learning image segmentation with cross-modality annotations</title><title>Medical image analysis</title><addtitle>Med Image Anal</addtitle><description>•We propose a novel cross-modality medical image segmentation method.•Our model can perform segmentation for a target domain without labeled training data.•A diverse data augmentation approach is used to augment the training data for segmentation.•Experiments in two different tasks demonstrate the effectiveness of proposed method. [Display omitted] The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.</description><subject>Annotations</subject><subject>Artificial neural networks</subject><subject>Computed tomography</subject><subject>Data augmentation</subject><subject>Disentangled representation learning</subject><subject>Generative adversarial learning</subject><subject>Generative adversarial networks</subject><subject>Image processing</subject><subject>Image segmentation</subject><subject>Magnetic resonance imaging</subject><subject>Medical image segmentation</subject><subject>Medical imaging</subject><subject>Neural networks</subject><issn>1361-8415</issn><issn>1361-8423</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><recordid>eNp9kU9P3DAQxa2qqGyBT1CpitQLlyz-n-QAUgUtIK3EBQ6cLCeeLF4lNtjJVnx7vBu6Ag6cbPn95nlmHkI_CJ4TTOTJat6DsXpOMSXphWKJv6AZYZLkJafs6-5OxD76HuMKY1xwjr-hfcYqUQhRztD9hV1DiJAZPehMj8se3KAH613W-pB1oIOzbpnZXi8hi_BG_2eHh6wJPsa890Z3dnjOtHN-kuMh2mt1F-Ho9TxAd3__3J5f5Yuby-vz34u84WUx5KYSLQOCOS6M4RQqoAVppKiF0djwohZtbYCArAEIJZXgEouqKHml05TasAN0Nvk-jnVaSJP6C7pTjyG1HJ6V11a9V5x9UEu_ViUpk1eVDI5fDYJ_GiEOqrexga7TDvwYFRWUM0GI5An99QFd-TG4NF6iJOOFTGCi2ERtlxOg3TVDsNpEp1ZqG53aRKem6FLVz7dz7Gr-Z5WA0wmAtM21haBiY8E1ySlAMyjj7acfvAAfvKzO</recordid><startdate>20210701</startdate><enddate>20210701</enddate><creator>Chen, Xu</creator><creator>Lian, Chunfeng</creator><creator>Wang, Li</creator><creator>Deng, Hannah</creator><creator>Kuang, Tianshu</creator><creator>Fung, Steve H.</creator><creator>Gateno, Jaime</creator><creator>Shen, Dinggang</creator><creator>Xia, James J.</creator><creator>Yap, Pew-Thian</creator><general>Elsevier B.V</general><general>Elsevier BV</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7QO</scope><scope>8FD</scope><scope>FR3</scope><scope>K9.</scope><scope>NAPCQ</scope><scope>P64</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0001-9363-1722</orcidid><orcidid>https://orcid.org/0000-0002-0367-3003</orcidid><orcidid>https://orcid.org/0000-0002-1177-682X</orcidid><orcidid>https://orcid.org/0000-0002-9319-6633</orcidid><orcidid>https://orcid.org/0000-0003-1489-2102</orcidid></search><sort><creationdate>20210701</creationdate><title>Diverse data augmentation for learning image segmentation with cross-modality annotations</title><author>Chen, Xu ; Lian, Chunfeng ; Wang, Li ; Deng, Hannah ; Kuang, Tianshu ; Fung, Steve H. ; Gateno, Jaime ; Shen, Dinggang ; Xia, James J. ; Yap, Pew-Thian</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c487t-d95f3e10407dd42e9e271c65b5da0d47b5fbde1e6bee12195460597849a842ad3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Annotations</topic><topic>Artificial neural networks</topic><topic>Computed tomography</topic><topic>Data augmentation</topic><topic>Disentangled representation learning</topic><topic>Generative adversarial learning</topic><topic>Generative adversarial networks</topic><topic>Image processing</topic><topic>Image segmentation</topic><topic>Magnetic resonance imaging</topic><topic>Medical image segmentation</topic><topic>Medical imaging</topic><topic>Neural networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Chen, Xu</creatorcontrib><creatorcontrib>Lian, Chunfeng</creatorcontrib><creatorcontrib>Wang, Li</creatorcontrib><creatorcontrib>Deng, Hannah</creatorcontrib><creatorcontrib>Kuang, Tianshu</creatorcontrib><creatorcontrib>Fung, Steve H.</creatorcontrib><creatorcontrib>Gateno, Jaime</creatorcontrib><creatorcontrib>Shen, Dinggang</creatorcontrib><creatorcontrib>Xia, James J.</creatorcontrib><creatorcontrib>Yap, Pew-Thian</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Biotechnology Research Abstracts</collection><collection>Technology Research Database</collection><collection>Engineering Research Database</collection><collection>ProQuest Health &amp; Medical Complete (Alumni)</collection><collection>Nursing &amp; Allied Health Premium</collection><collection>Biotechnology and BioEngineering Abstracts</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Medical image analysis</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Chen, Xu</au><au>Lian, Chunfeng</au><au>Wang, Li</au><au>Deng, Hannah</au><au>Kuang, Tianshu</au><au>Fung, Steve H.</au><au>Gateno, Jaime</au><au>Shen, Dinggang</au><au>Xia, James J.</au><au>Yap, Pew-Thian</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Diverse data augmentation for learning image segmentation with cross-modality annotations</atitle><jtitle>Medical image analysis</jtitle><addtitle>Med Image Anal</addtitle><date>2021-07-01</date><risdate>2021</risdate><volume>71</volume><spage>102060</spage><epage>102060</epage><pages>102060-102060</pages><artnum>102060</artnum><issn>1361-8415</issn><eissn>1361-8423</eissn><abstract>•We propose a novel cross-modality medical image segmentation method.•Our model can perform segmentation for a target domain without labeled training data.•A diverse data augmentation approach is used to augment the training data for segmentation.•Experiments in two different tasks demonstrate the effectiveness of proposed method. [Display omitted] The dearth of annotated data is a major hurdle in building reliable image segmentation models. Manual annotation of medical images is tedious, time-consuming, and significantly variable across imaging modalities. The need for annotation can be ameliorated by leveraging an annotation-rich source modality in learning a segmentation model for an annotation-poor target modality. In this paper, we introduce a diverse data augmentation generative adversarial network (DDA-GAN) to train a segmentation model for an unannotated target image domain by borrowing information from an annotated source image domain. This is achieved by generating diverse augmented data for the target domain by one-to-many source-to-target translation. The DDA-GAN uses unpaired images from the source and target domains and is an end-to-end convolutional neural network that (i) explicitly disentangles domain-invariant structural features related to segmentation from domain-specific appearance features, (ii) combines structural features from the source domain with appearance features randomly sampled from the target domain for data augmentation, and (iii) train the segmentation model with the augmented data in the target domain and the annotations from the source domain. The effectiveness of our method is demonstrated both qualitatively and quantitatively in comparison with the state of the art for segmentation of craniomaxillofacial bony structures via MRI and cardiac substructures via CT.</abstract><cop>Netherlands</cop><pub>Elsevier B.V</pub><pmid>33957558</pmid><doi>10.1016/j.media.2021.102060</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0001-9363-1722</orcidid><orcidid>https://orcid.org/0000-0002-0367-3003</orcidid><orcidid>https://orcid.org/0000-0002-1177-682X</orcidid><orcidid>https://orcid.org/0000-0002-9319-6633</orcidid><orcidid>https://orcid.org/0000-0003-1489-2102</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1361-8415
ispartof Medical image analysis, 2021-07, Vol.71, p.102060-102060, Article 102060
issn 1361-8415
1361-8423
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_8184609
source Elsevier ScienceDirect Journals Complete
subjects Annotations
Artificial neural networks
Computed tomography
Data augmentation
Disentangled representation learning
Generative adversarial learning
Generative adversarial networks
Image processing
Image segmentation
Magnetic resonance imaging
Medical image segmentation
Medical imaging
Neural networks
title Diverse data augmentation for learning image segmentation with cross-modality annotations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-13T20%3A16%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Diverse%20data%20augmentation%20for%20learning%20image%20segmentation%20with%20cross-modality%20annotations&rft.jtitle=Medical%20image%20analysis&rft.au=Chen,%20Xu&rft.date=2021-07-01&rft.volume=71&rft.spage=102060&rft.epage=102060&rft.pages=102060-102060&rft.artnum=102060&rft.issn=1361-8415&rft.eissn=1361-8423&rft_id=info:doi/10.1016/j.media.2021.102060&rft_dat=%3Cproquest_pubme%3E2524351164%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2563476351&rft_id=info:pmid/33957558&rft_els_id=S1361841521001067&rfr_iscdi=true