Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation
U-Net is a deep learning technique that has made significant contributions to medical image segmentation. Although the accomplishments of deep learning algorithms in terms of image processing are evident, many challenges still need to be overcome to achieve human-like performance. One of the main ch...
Gespeichert in:
Veröffentlicht in: | Journal of medical imaging (Bellingham, Wash.) Wash.), 2022-11, Vol.9 (6), p.064004-064004 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 064004 |
---|---|
container_issue | 6 |
container_start_page | 064004 |
container_title | Journal of medical imaging (Bellingham, Wash.) |
container_volume | 9 |
creator | Siddique, Nahian Paheding, Sidike Reyes Angulo, Abel A. Alom, Md. Zahangir Devabhaktuni, Vijay K. |
description | U-Net is a deep learning technique that has made significant contributions to medical image segmentation. Although the accomplishments of deep learning algorithms in terms of image processing are evident, many challenges still need to be overcome to achieve human-like performance. One of the main challenges in building deeper U-Nets is black-box problems, such as vanishing gradients. Overcoming this problem allows us to develop neural networks with advanced network designs.
We propose three U-Net variants, namely efficient R2U-Net, efficient dense U-Net, and efficient fractal U-Net, that can create highly accurate segmentation maps. The first part of our contribution makes use of EfficientNet to distribute resources in the network efficiently. The second part of our work applies the following layer connections to design the U-Net decoders: residual connections, dense connections, and fractal expansion. We apply EfficientNet as the encoder to our three decoders to design three conceivable models.
The aforementioned three proposed deep learning models were tested on four benchmark datasets, including the CHASE DB1 and digital retinal images for vessel extraction (DRIVE) retinal image databases and the ISIC 2018 and HAM10000 dermoscopy image databases. We obtained the highest Dice coefficient of 0.8013, 0.8808, 0.8019, and 0.9295 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively, and a Jaccard (JAC) score of 0.6686, 0.7870, 0.6694, and 0.8683 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively. Statistical analysis revealed that the proposed deep learning models achieved better segmentation results compared with the state-of-the-art models.
U-Net is quite an adaptable deep learning framework and can be integrated with other deep learning techniques. The use of recurrent feedback connections, dense convolution, residual skip connections, and fractal convolutional expansions allow for the design of improved deeper U-Net models. With the addition of EfficientNet, we can now leverage the performance of an optimally scaled classifier for U-Net encoders. |
doi_str_mv | 10.1117/1.JMI.9.6.064004 |
format | Article |
fullrecord | <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9789743</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2760171152</sourcerecordid><originalsourceid>FETCH-LOGICAL-c392t-816f26b58fbd01656b4e60d9a1bc7b036ceb6eabb19f9ba7c4047c5a032caa223</originalsourceid><addsrcrecordid>eNp1kUFv1DAQhS0EolXpnRPykUMTZpzEWV-QUNVCUYELPVu2M9n1KhsvtlPEv8erLSs4cLKl-ea9p3mMvUaoEbF_h_XnL3e1qmUNsgVon7Fz0QhVtQ3C89MfxBm7TGkLAIjQCWxfsrNGdgoliHO2vY3GZTNd8UhuiZHmfMXNPPCB5kT8ofpKmZvoNj6Ty0ukxH_6vOE34-idL_RhTrMLA0U-hsh3NHhnJu53Zk080XpXIJN9mF-xF6OZEl0-vRfs4fbm-_Wn6v7bx7vrD_eVa5TI1QrlKKTtVqMdAGUnbUsSBmXQut5CIx1ZScZaVKOypncttL3rDDTCGSNEc8HeH3X3iy1pXPGPZtL7WCLFXzoYr_-dzH6j1-FRq36l-rYpAm-fBGL4sVDKeueTo2kyM4UladFLwB6xO3jBEXUxpBRpPNkg6ENLGnVpSSst9bGlsvLm73inhT-dFKA6AmnvSW_DEudyrv8L_gan3p3R</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2760171152</pqid></control><display><type>article</type><title>Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation</title><source>EZB-FREE-00999 freely available EZB journals</source><source>PubMed Central</source><creator>Siddique, Nahian ; Paheding, Sidike ; Reyes Angulo, Abel A. ; Alom, Md. Zahangir ; Devabhaktuni, Vijay K.</creator><creatorcontrib>Siddique, Nahian ; Paheding, Sidike ; Reyes Angulo, Abel A. ; Alom, Md. Zahangir ; Devabhaktuni, Vijay K.</creatorcontrib><description>U-Net is a deep learning technique that has made significant contributions to medical image segmentation. Although the accomplishments of deep learning algorithms in terms of image processing are evident, many challenges still need to be overcome to achieve human-like performance. One of the main challenges in building deeper U-Nets is black-box problems, such as vanishing gradients. Overcoming this problem allows us to develop neural networks with advanced network designs.
We propose three U-Net variants, namely efficient R2U-Net, efficient dense U-Net, and efficient fractal U-Net, that can create highly accurate segmentation maps. The first part of our contribution makes use of EfficientNet to distribute resources in the network efficiently. The second part of our work applies the following layer connections to design the U-Net decoders: residual connections, dense connections, and fractal expansion. We apply EfficientNet as the encoder to our three decoders to design three conceivable models.
The aforementioned three proposed deep learning models were tested on four benchmark datasets, including the CHASE DB1 and digital retinal images for vessel extraction (DRIVE) retinal image databases and the ISIC 2018 and HAM10000 dermoscopy image databases. We obtained the highest Dice coefficient of 0.8013, 0.8808, 0.8019, and 0.9295 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively, and a Jaccard (JAC) score of 0.6686, 0.7870, 0.6694, and 0.8683 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively. Statistical analysis revealed that the proposed deep learning models achieved better segmentation results compared with the state-of-the-art models.
U-Net is quite an adaptable deep learning framework and can be integrated with other deep learning techniques. The use of recurrent feedback connections, dense convolution, residual skip connections, and fractal convolutional expansions allow for the design of improved deeper U-Net models. With the addition of EfficientNet, we can now leverage the performance of an optimally scaled classifier for U-Net encoders.</description><identifier>ISSN: 2329-4302</identifier><identifier>EISSN: 2329-4310</identifier><identifier>DOI: 10.1117/1.JMI.9.6.064004</identifier><identifier>PMID: 36591602</identifier><language>eng</language><publisher>United States: Society of Photo-Optical Instrumentation Engineers</publisher><subject>Image Processing</subject><ispartof>Journal of medical imaging (Bellingham, Wash.), 2022-11, Vol.9 (6), p.064004-064004</ispartof><rights>2022 Society of Photo-Optical Instrumentation Engineers (SPIE)</rights><rights>2022 Society of Photo-Optical Instrumentation Engineers (SPIE).</rights><rights>2022 Society of Photo-Optical Instrumentation Engineers (SPIE) 2022 Society of Photo-Optical Instrumentation Engineers</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c392t-816f26b58fbd01656b4e60d9a1bc7b036ceb6eabb19f9ba7c4047c5a032caa223</citedby><cites>FETCH-LOGICAL-c392t-816f26b58fbd01656b4e60d9a1bc7b036ceb6eabb19f9ba7c4047c5a032caa223</cites><orcidid>0000-0003-0332-8231 ; 0000-0003-4712-9672 ; 0000-0002-2314-1207</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9789743/pdf/$$EPDF$$P50$$Gpubmedcentral$$H</linktopdf><linktohtml>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9789743/$$EHTML$$P50$$Gpubmedcentral$$H</linktohtml><link.rule.ids>230,314,727,780,784,885,27922,27923,53789,53791</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36591602$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Siddique, Nahian</creatorcontrib><creatorcontrib>Paheding, Sidike</creatorcontrib><creatorcontrib>Reyes Angulo, Abel A.</creatorcontrib><creatorcontrib>Alom, Md. Zahangir</creatorcontrib><creatorcontrib>Devabhaktuni, Vijay K.</creatorcontrib><title>Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation</title><title>Journal of medical imaging (Bellingham, Wash.)</title><addtitle>J. Med. Imag</addtitle><description>U-Net is a deep learning technique that has made significant contributions to medical image segmentation. Although the accomplishments of deep learning algorithms in terms of image processing are evident, many challenges still need to be overcome to achieve human-like performance. One of the main challenges in building deeper U-Nets is black-box problems, such as vanishing gradients. Overcoming this problem allows us to develop neural networks with advanced network designs.
We propose three U-Net variants, namely efficient R2U-Net, efficient dense U-Net, and efficient fractal U-Net, that can create highly accurate segmentation maps. The first part of our contribution makes use of EfficientNet to distribute resources in the network efficiently. The second part of our work applies the following layer connections to design the U-Net decoders: residual connections, dense connections, and fractal expansion. We apply EfficientNet as the encoder to our three decoders to design three conceivable models.
The aforementioned three proposed deep learning models were tested on four benchmark datasets, including the CHASE DB1 and digital retinal images for vessel extraction (DRIVE) retinal image databases and the ISIC 2018 and HAM10000 dermoscopy image databases. We obtained the highest Dice coefficient of 0.8013, 0.8808, 0.8019, and 0.9295 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively, and a Jaccard (JAC) score of 0.6686, 0.7870, 0.6694, and 0.8683 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively. Statistical analysis revealed that the proposed deep learning models achieved better segmentation results compared with the state-of-the-art models.
U-Net is quite an adaptable deep learning framework and can be integrated with other deep learning techniques. The use of recurrent feedback connections, dense convolution, residual skip connections, and fractal convolutional expansions allow for the design of improved deeper U-Net models. With the addition of EfficientNet, we can now leverage the performance of an optimally scaled classifier for U-Net encoders.</description><subject>Image Processing</subject><issn>2329-4302</issn><issn>2329-4310</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><recordid>eNp1kUFv1DAQhS0EolXpnRPykUMTZpzEWV-QUNVCUYELPVu2M9n1KhsvtlPEv8erLSs4cLKl-ea9p3mMvUaoEbF_h_XnL3e1qmUNsgVon7Fz0QhVtQ3C89MfxBm7TGkLAIjQCWxfsrNGdgoliHO2vY3GZTNd8UhuiZHmfMXNPPCB5kT8ofpKmZvoNj6Ty0ukxH_6vOE34-idL_RhTrMLA0U-hsh3NHhnJu53Zk080XpXIJN9mF-xF6OZEl0-vRfs4fbm-_Wn6v7bx7vrD_eVa5TI1QrlKKTtVqMdAGUnbUsSBmXQut5CIx1ZScZaVKOypncttL3rDDTCGSNEc8HeH3X3iy1pXPGPZtL7WCLFXzoYr_-dzH6j1-FRq36l-rYpAm-fBGL4sVDKeueTo2kyM4UladFLwB6xO3jBEXUxpBRpPNkg6ENLGnVpSSst9bGlsvLm73inhT-dFKA6AmnvSW_DEudyrv8L_gan3p3R</recordid><startdate>20221101</startdate><enddate>20221101</enddate><creator>Siddique, Nahian</creator><creator>Paheding, Sidike</creator><creator>Reyes Angulo, Abel A.</creator><creator>Alom, Md. Zahangir</creator><creator>Devabhaktuni, Vijay K.</creator><general>Society of Photo-Optical Instrumentation Engineers</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope><orcidid>https://orcid.org/0000-0003-0332-8231</orcidid><orcidid>https://orcid.org/0000-0003-4712-9672</orcidid><orcidid>https://orcid.org/0000-0002-2314-1207</orcidid></search><sort><creationdate>20221101</creationdate><title>Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation</title><author>Siddique, Nahian ; Paheding, Sidike ; Reyes Angulo, Abel A. ; Alom, Md. Zahangir ; Devabhaktuni, Vijay K.</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c392t-816f26b58fbd01656b4e60d9a1bc7b036ceb6eabb19f9ba7c4047c5a032caa223</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Image Processing</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Siddique, Nahian</creatorcontrib><creatorcontrib>Paheding, Sidike</creatorcontrib><creatorcontrib>Reyes Angulo, Abel A.</creatorcontrib><creatorcontrib>Alom, Md. Zahangir</creatorcontrib><creatorcontrib>Devabhaktuni, Vijay K.</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Journal of medical imaging (Bellingham, Wash.)</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Siddique, Nahian</au><au>Paheding, Sidike</au><au>Reyes Angulo, Abel A.</au><au>Alom, Md. Zahangir</au><au>Devabhaktuni, Vijay K.</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation</atitle><jtitle>Journal of medical imaging (Bellingham, Wash.)</jtitle><addtitle>J. Med. Imag</addtitle><date>2022-11-01</date><risdate>2022</risdate><volume>9</volume><issue>6</issue><spage>064004</spage><epage>064004</epage><pages>064004-064004</pages><issn>2329-4302</issn><eissn>2329-4310</eissn><abstract>U-Net is a deep learning technique that has made significant contributions to medical image segmentation. Although the accomplishments of deep learning algorithms in terms of image processing are evident, many challenges still need to be overcome to achieve human-like performance. One of the main challenges in building deeper U-Nets is black-box problems, such as vanishing gradients. Overcoming this problem allows us to develop neural networks with advanced network designs.
We propose three U-Net variants, namely efficient R2U-Net, efficient dense U-Net, and efficient fractal U-Net, that can create highly accurate segmentation maps. The first part of our contribution makes use of EfficientNet to distribute resources in the network efficiently. The second part of our work applies the following layer connections to design the U-Net decoders: residual connections, dense connections, and fractal expansion. We apply EfficientNet as the encoder to our three decoders to design three conceivable models.
The aforementioned three proposed deep learning models were tested on four benchmark datasets, including the CHASE DB1 and digital retinal images for vessel extraction (DRIVE) retinal image databases and the ISIC 2018 and HAM10000 dermoscopy image databases. We obtained the highest Dice coefficient of 0.8013, 0.8808, 0.8019, and 0.9295 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively, and a Jaccard (JAC) score of 0.6686, 0.7870, 0.6694, and 0.8683 for CHASE DB1, ISIC 2018, DRIVE, and HAM10000, respectively. Statistical analysis revealed that the proposed deep learning models achieved better segmentation results compared with the state-of-the-art models.
U-Net is quite an adaptable deep learning framework and can be integrated with other deep learning techniques. The use of recurrent feedback connections, dense convolution, residual skip connections, and fractal convolutional expansions allow for the design of improved deeper U-Net models. With the addition of EfficientNet, we can now leverage the performance of an optimally scaled classifier for U-Net encoders.</abstract><cop>United States</cop><pub>Society of Photo-Optical Instrumentation Engineers</pub><pmid>36591602</pmid><doi>10.1117/1.JMI.9.6.064004</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0003-0332-8231</orcidid><orcidid>https://orcid.org/0000-0003-4712-9672</orcidid><orcidid>https://orcid.org/0000-0002-2314-1207</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2329-4302 |
ispartof | Journal of medical imaging (Bellingham, Wash.), 2022-11, Vol.9 (6), p.064004-064004 |
issn | 2329-4302 2329-4310 |
language | eng |
recordid | cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9789743 |
source | EZB-FREE-00999 freely available EZB journals; PubMed Central |
subjects | Image Processing |
title | Fractal, recurrent, and dense U-Net architectures with EfficientNet encoder for medical image segmentation |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-10T07%3A02%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Fractal,%20recurrent,%20and%20dense%20U-Net%20architectures%20with%20EfficientNet%20encoder%20for%20medical%20image%20segmentation&rft.jtitle=Journal%20of%20medical%20imaging%20(Bellingham,%20Wash.)&rft.au=Siddique,%20Nahian&rft.date=2022-11-01&rft.volume=9&rft.issue=6&rft.spage=064004&rft.epage=064004&rft.pages=064004-064004&rft.issn=2329-4302&rft.eissn=2329-4310&rft_id=info:doi/10.1117/1.JMI.9.6.064004&rft_dat=%3Cproquest_pubme%3E2760171152%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2760171152&rft_id=info:pmid/36591602&rfr_iscdi=true |