Light mixed‐supervised segmentation for 3D medical image data

Background Accurate 3D semantic segmentation models are essential for many clinical applications. To train a model for 3D segmentation, voxel‐level annotation is necessary, which is expensive to obtain due to laborious work and privacy protection. To accurately annotate 3D medical data, such as MRI,...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Medical physics (Lancaster) 2024-01, Vol.51 (1), p.167-178
Hauptverfasser: Yang, Hongxu, Tan, Tao, Tegzes, Pal, Dong, Xiaomeng, Tamada, Rajesh, Ferenczi, Lehel, Avinash, Gopal
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 178
container_issue 1
container_start_page 167
container_title Medical physics (Lancaster)
container_volume 51
creator Yang, Hongxu
Tan, Tao
Tegzes, Pal
Dong, Xiaomeng
Tamada, Rajesh
Ferenczi, Lehel
Avinash, Gopal
description Background Accurate 3D semantic segmentation models are essential for many clinical applications. To train a model for 3D segmentation, voxel‐level annotation is necessary, which is expensive to obtain due to laborious work and privacy protection. To accurately annotate 3D medical data, such as MRI, a common practice is to annotate the volumetric data in a slice‐by‐slice contouring way along principal axes. Purpose In order to reduce the annotation effort in slices, weakly supervised learning with a bounding box (Bbox) was proposed to leverage the discriminating information via a tightness prior assumption. Nevertheless, this method requests accurate and tight Bboxes, which will significantly drop the performance when tightness is not held, that is when a relaxed Bbox is applied. Therefore, there is a need to train a stable model based on relaxed Bbox annotation. Methods This paper presents a mixed‐supervised training strategy to reduce the annotation effort for 3D segmentation tasks. In the proposed approach, a fully annotated contour is only required for a single slice of the volume. In contrast, the rest of the slices with targets are annotated with relaxed Bboxes. This mixed‐supervised method adopts fully supervised learning, relaxed Bbox prior, and contrastive learning during the training, which ensures the network exploits the discriminative information of the training volumes properly. The proposed method was evaluated on two public 3D medical imaging datasets (MRI prostate dataset and Vestibular Schwannoma [VS] dataset). Results The proposed method obtained a high segmentation Dice score of 85.3% on an MRI prostate dataset and 83.3% on a VS dataset with relaxed Bbox annotation, which are close to a fully supervised model. Moreover, with the same relaxed Bbox annotations, the proposed method outperforms the state‐of‐the‐art methods. More importantly, the model performance is stable when the accuracy of Bbox annotation varies. Conclusions The presented study proposes a method based on a mixed‐supervised learning method in 3D medical imaging. The benefit will be stable segmentation of the target in 3D images with low accurate annotation requirement, which leads to easier model training on large‐scale datasets.
doi_str_mv 10.1002/mp.16816
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_2885203138</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2885203138</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2826-6b1d3f33f8f312bcfeebb7bc011e743bddeaa86476dd3756764c0305bde7d7003</originalsourceid><addsrcrecordid>eNp1kL1OwzAURi0EoqUg8QQoI0vKtZ3Y7oRQ-ZWKYIDZsuObYpQ0IU6AbjwCz8iTEGiBiekuR0ffPYTsUxhTAHZU1mMqFBUbZMgSyeOEwWSTDAEmScwSSAdkJ4RHABA8hW0y4HICE8X5kBzP_PyhjUr_iu7j7T10NTbPPqCLAs5LXLSm9dUiyqsm4qdRic5npoh8aeYYOdOaXbKVmyLg3vqOyP352d30Mp7dXFxNT2ZxxhQTsbDU8ZzzXOWcMpvliNZKmwGlKBNunUNjlEikcI7LVEiRZMAhtQ6lkwB8RA5X3rqpnjoMrS59yLAozAKrLmimVMqAU67-0KypQmgw13XTD26WmoL-yqXLWn_n6tGDtbWz_W-_4E-fHohXwIsvcPmvSF_froSfgX5zhw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2885203138</pqid></control><display><type>article</type><title>Light mixed‐supervised segmentation for 3D medical image data</title><source>MEDLINE</source><source>Wiley Online Library Journals Frontfile Complete</source><source>Alma/SFX Local Collection</source><creator>Yang, Hongxu ; Tan, Tao ; Tegzes, Pal ; Dong, Xiaomeng ; Tamada, Rajesh ; Ferenczi, Lehel ; Avinash, Gopal</creator><creatorcontrib>Yang, Hongxu ; Tan, Tao ; Tegzes, Pal ; Dong, Xiaomeng ; Tamada, Rajesh ; Ferenczi, Lehel ; Avinash, Gopal</creatorcontrib><description>Background Accurate 3D semantic segmentation models are essential for many clinical applications. To train a model for 3D segmentation, voxel‐level annotation is necessary, which is expensive to obtain due to laborious work and privacy protection. To accurately annotate 3D medical data, such as MRI, a common practice is to annotate the volumetric data in a slice‐by‐slice contouring way along principal axes. Purpose In order to reduce the annotation effort in slices, weakly supervised learning with a bounding box (Bbox) was proposed to leverage the discriminating information via a tightness prior assumption. Nevertheless, this method requests accurate and tight Bboxes, which will significantly drop the performance when tightness is not held, that is when a relaxed Bbox is applied. Therefore, there is a need to train a stable model based on relaxed Bbox annotation. Methods This paper presents a mixed‐supervised training strategy to reduce the annotation effort for 3D segmentation tasks. In the proposed approach, a fully annotated contour is only required for a single slice of the volume. In contrast, the rest of the slices with targets are annotated with relaxed Bboxes. This mixed‐supervised method adopts fully supervised learning, relaxed Bbox prior, and contrastive learning during the training, which ensures the network exploits the discriminative information of the training volumes properly. The proposed method was evaluated on two public 3D medical imaging datasets (MRI prostate dataset and Vestibular Schwannoma [VS] dataset). Results The proposed method obtained a high segmentation Dice score of 85.3% on an MRI prostate dataset and 83.3% on a VS dataset with relaxed Bbox annotation, which are close to a fully supervised model. Moreover, with the same relaxed Bbox annotations, the proposed method outperforms the state‐of‐the‐art methods. More importantly, the model performance is stable when the accuracy of Bbox annotation varies. Conclusions The presented study proposes a method based on a mixed‐supervised learning method in 3D medical imaging. The benefit will be stable segmentation of the target in 3D images with low accurate annotation requirement, which leads to easier model training on large‐scale datasets.</description><identifier>ISSN: 0094-2405</identifier><identifier>EISSN: 2473-4209</identifier><identifier>DOI: 10.1002/mp.16816</identifier><identifier>PMID: 37909833</identifier><language>eng</language><publisher>United States</publisher><subject>3D medical images ; contrastive learning ; Humans ; Image Processing, Computer-Assisted ; Imaging, Three-Dimensional ; Male ; mixed‐supervised learning ; Neuroma, Acoustic ; Pelvis ; Prostate ; relaxed bounding box ; Supervised Machine Learning</subject><ispartof>Medical physics (Lancaster), 2024-01, Vol.51 (1), p.167-178</ispartof><rights>2023 American Association of Physicists in Medicine.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c2826-6b1d3f33f8f312bcfeebb7bc011e743bddeaa86476dd3756764c0305bde7d7003</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://onlinelibrary.wiley.com/doi/pdf/10.1002%2Fmp.16816$$EPDF$$P50$$Gwiley$$H</linktopdf><linktohtml>$$Uhttps://onlinelibrary.wiley.com/doi/full/10.1002%2Fmp.16816$$EHTML$$P50$$Gwiley$$H</linktohtml><link.rule.ids>314,776,780,1411,27901,27902,45550,45551</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37909833$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Yang, Hongxu</creatorcontrib><creatorcontrib>Tan, Tao</creatorcontrib><creatorcontrib>Tegzes, Pal</creatorcontrib><creatorcontrib>Dong, Xiaomeng</creatorcontrib><creatorcontrib>Tamada, Rajesh</creatorcontrib><creatorcontrib>Ferenczi, Lehel</creatorcontrib><creatorcontrib>Avinash, Gopal</creatorcontrib><title>Light mixed‐supervised segmentation for 3D medical image data</title><title>Medical physics (Lancaster)</title><addtitle>Med Phys</addtitle><description>Background Accurate 3D semantic segmentation models are essential for many clinical applications. To train a model for 3D segmentation, voxel‐level annotation is necessary, which is expensive to obtain due to laborious work and privacy protection. To accurately annotate 3D medical data, such as MRI, a common practice is to annotate the volumetric data in a slice‐by‐slice contouring way along principal axes. Purpose In order to reduce the annotation effort in slices, weakly supervised learning with a bounding box (Bbox) was proposed to leverage the discriminating information via a tightness prior assumption. Nevertheless, this method requests accurate and tight Bboxes, which will significantly drop the performance when tightness is not held, that is when a relaxed Bbox is applied. Therefore, there is a need to train a stable model based on relaxed Bbox annotation. Methods This paper presents a mixed‐supervised training strategy to reduce the annotation effort for 3D segmentation tasks. In the proposed approach, a fully annotated contour is only required for a single slice of the volume. In contrast, the rest of the slices with targets are annotated with relaxed Bboxes. This mixed‐supervised method adopts fully supervised learning, relaxed Bbox prior, and contrastive learning during the training, which ensures the network exploits the discriminative information of the training volumes properly. The proposed method was evaluated on two public 3D medical imaging datasets (MRI prostate dataset and Vestibular Schwannoma [VS] dataset). Results The proposed method obtained a high segmentation Dice score of 85.3% on an MRI prostate dataset and 83.3% on a VS dataset with relaxed Bbox annotation, which are close to a fully supervised model. Moreover, with the same relaxed Bbox annotations, the proposed method outperforms the state‐of‐the‐art methods. More importantly, the model performance is stable when the accuracy of Bbox annotation varies. Conclusions The presented study proposes a method based on a mixed‐supervised learning method in 3D medical imaging. The benefit will be stable segmentation of the target in 3D images with low accurate annotation requirement, which leads to easier model training on large‐scale datasets.</description><subject>3D medical images</subject><subject>contrastive learning</subject><subject>Humans</subject><subject>Image Processing, Computer-Assisted</subject><subject>Imaging, Three-Dimensional</subject><subject>Male</subject><subject>mixed‐supervised learning</subject><subject>Neuroma, Acoustic</subject><subject>Pelvis</subject><subject>Prostate</subject><subject>relaxed bounding box</subject><subject>Supervised Machine Learning</subject><issn>0094-2405</issn><issn>2473-4209</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNp1kL1OwzAURi0EoqUg8QQoI0vKtZ3Y7oRQ-ZWKYIDZsuObYpQ0IU6AbjwCz8iTEGiBiekuR0ffPYTsUxhTAHZU1mMqFBUbZMgSyeOEwWSTDAEmScwSSAdkJ4RHABA8hW0y4HICE8X5kBzP_PyhjUr_iu7j7T10NTbPPqCLAs5LXLSm9dUiyqsm4qdRic5npoh8aeYYOdOaXbKVmyLg3vqOyP352d30Mp7dXFxNT2ZxxhQTsbDU8ZzzXOWcMpvliNZKmwGlKBNunUNjlEikcI7LVEiRZMAhtQ6lkwB8RA5X3rqpnjoMrS59yLAozAKrLmimVMqAU67-0KypQmgw13XTD26WmoL-yqXLWn_n6tGDtbWz_W-_4E-fHohXwIsvcPmvSF_froSfgX5zhw</recordid><startdate>202401</startdate><enddate>202401</enddate><creator>Yang, Hongxu</creator><creator>Tan, Tao</creator><creator>Tegzes, Pal</creator><creator>Dong, Xiaomeng</creator><creator>Tamada, Rajesh</creator><creator>Ferenczi, Lehel</creator><creator>Avinash, Gopal</creator><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope></search><sort><creationdate>202401</creationdate><title>Light mixed‐supervised segmentation for 3D medical image data</title><author>Yang, Hongxu ; Tan, Tao ; Tegzes, Pal ; Dong, Xiaomeng ; Tamada, Rajesh ; Ferenczi, Lehel ; Avinash, Gopal</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2826-6b1d3f33f8f312bcfeebb7bc011e743bddeaa86476dd3756764c0305bde7d7003</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>3D medical images</topic><topic>contrastive learning</topic><topic>Humans</topic><topic>Image Processing, Computer-Assisted</topic><topic>Imaging, Three-Dimensional</topic><topic>Male</topic><topic>mixed‐supervised learning</topic><topic>Neuroma, Acoustic</topic><topic>Pelvis</topic><topic>Prostate</topic><topic>relaxed bounding box</topic><topic>Supervised Machine Learning</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Yang, Hongxu</creatorcontrib><creatorcontrib>Tan, Tao</creatorcontrib><creatorcontrib>Tegzes, Pal</creatorcontrib><creatorcontrib>Dong, Xiaomeng</creatorcontrib><creatorcontrib>Tamada, Rajesh</creatorcontrib><creatorcontrib>Ferenczi, Lehel</creatorcontrib><creatorcontrib>Avinash, Gopal</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Medical physics (Lancaster)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, Hongxu</au><au>Tan, Tao</au><au>Tegzes, Pal</au><au>Dong, Xiaomeng</au><au>Tamada, Rajesh</au><au>Ferenczi, Lehel</au><au>Avinash, Gopal</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Light mixed‐supervised segmentation for 3D medical image data</atitle><jtitle>Medical physics (Lancaster)</jtitle><addtitle>Med Phys</addtitle><date>2024-01</date><risdate>2024</risdate><volume>51</volume><issue>1</issue><spage>167</spage><epage>178</epage><pages>167-178</pages><issn>0094-2405</issn><eissn>2473-4209</eissn><abstract>Background Accurate 3D semantic segmentation models are essential for many clinical applications. To train a model for 3D segmentation, voxel‐level annotation is necessary, which is expensive to obtain due to laborious work and privacy protection. To accurately annotate 3D medical data, such as MRI, a common practice is to annotate the volumetric data in a slice‐by‐slice contouring way along principal axes. Purpose In order to reduce the annotation effort in slices, weakly supervised learning with a bounding box (Bbox) was proposed to leverage the discriminating information via a tightness prior assumption. Nevertheless, this method requests accurate and tight Bboxes, which will significantly drop the performance when tightness is not held, that is when a relaxed Bbox is applied. Therefore, there is a need to train a stable model based on relaxed Bbox annotation. Methods This paper presents a mixed‐supervised training strategy to reduce the annotation effort for 3D segmentation tasks. In the proposed approach, a fully annotated contour is only required for a single slice of the volume. In contrast, the rest of the slices with targets are annotated with relaxed Bboxes. This mixed‐supervised method adopts fully supervised learning, relaxed Bbox prior, and contrastive learning during the training, which ensures the network exploits the discriminative information of the training volumes properly. The proposed method was evaluated on two public 3D medical imaging datasets (MRI prostate dataset and Vestibular Schwannoma [VS] dataset). Results The proposed method obtained a high segmentation Dice score of 85.3% on an MRI prostate dataset and 83.3% on a VS dataset with relaxed Bbox annotation, which are close to a fully supervised model. Moreover, with the same relaxed Bbox annotations, the proposed method outperforms the state‐of‐the‐art methods. More importantly, the model performance is stable when the accuracy of Bbox annotation varies. Conclusions The presented study proposes a method based on a mixed‐supervised learning method in 3D medical imaging. The benefit will be stable segmentation of the target in 3D images with low accurate annotation requirement, which leads to easier model training on large‐scale datasets.</abstract><cop>United States</cop><pmid>37909833</pmid><doi>10.1002/mp.16816</doi><tpages>12</tpages></addata></record>
fulltext fulltext
identifier ISSN: 0094-2405
ispartof Medical physics (Lancaster), 2024-01, Vol.51 (1), p.167-178
issn 0094-2405
2473-4209
language eng
recordid cdi_proquest_miscellaneous_2885203138
source MEDLINE; Wiley Online Library Journals Frontfile Complete; Alma/SFX Local Collection
subjects 3D medical images
contrastive learning
Humans
Image Processing, Computer-Assisted
Imaging, Three-Dimensional
Male
mixed‐supervised learning
Neuroma, Acoustic
Pelvis
Prostate
relaxed bounding box
Supervised Machine Learning
title Light mixed‐supervised segmentation for 3D medical image data
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T02%3A36%3A09IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Light%20mixed%E2%80%90supervised%20segmentation%20for%203D%20medical%20image%20data&rft.jtitle=Medical%20physics%20(Lancaster)&rft.au=Yang,%20Hongxu&rft.date=2024-01&rft.volume=51&rft.issue=1&rft.spage=167&rft.epage=178&rft.pages=167-178&rft.issn=0094-2405&rft.eissn=2473-4209&rft_id=info:doi/10.1002/mp.16816&rft_dat=%3Cproquest_cross%3E2885203138%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2885203138&rft_id=info:pmid/37909833&rfr_iscdi=true