Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation

The detailed physiological perspectives captured by medical imaging provides actionable insights to doctors to manage comprehensive care of patients. However, the quality of such diagnostic image modalities is often affected by mismanagement of the image capturing process by poorly trained technicia...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Biomedical physics & engineering express 2021-11, Vol.8 (1), p.15011
Hauptverfasser: K, Karthik, Kamath S, Sowmya
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 1
container_start_page 15011
container_title Biomedical physics & engineering express
container_volume 8
creator K, Karthik
Kamath S, Sowmya
description The detailed physiological perspectives captured by medical imaging provides actionable insights to doctors to manage comprehensive care of patients. However, the quality of such diagnostic image modalities is often affected by mismanagement of the image capturing process by poorly trained technicians and older/poorly maintained imaging equipment. Further, a patient is often subjected to scanning at different orientations to capture the frontal, lateral and sagittal views of the affected areas. Due to the large volume of diagnostic scans performed at a modern hospital, adequate documentation of such additional perspectives is mostly overlooked, which is also an essential key element of quality diagnostic systems and predictive analytics systems. Another crucial challenge affecting effective medical image data management is that the diagnostic scans are essentially stored as unstructured data, lacking a well-defined processing methodology for enabling intelligent image data management for supporting applications like similar patient retrieval , automated disease prediction etc. One solution is to incorporate automated diagnostic image descriptions of the observation/findings by leveraging computer vision and natural language processing. In this work, we present multi-task neural models capable of addressing these critical challenges. We propose ESRGAN, an image enhancement technique for improving the quality and visualization of medical chest x-ray images, thereby substantially improving the potential for accurate diagnosis, automatic detection and region-of-interest segmentation. We also propose a CNN-based model called ViewNet for predicting the view orientation of the x-ray image and generating a medical report using Xception net, thus facilitating a robust medical image management system for intelligent diagnosis applications. Experimental results are demonstrated using standard metrics like BRISQUE, PIQE and BLEU scores, indicating that the proposed models achieved excellent performance. Further, the proposed deep learning approaches enable diagnosis in a lesser time and their hybrid architecture shows significant potential for supporting many intelligent diagnosis applications.
doi_str_mv 10.1088/2057-1976/ac3add
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmed_primary_34789588</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2599075296</sourcerecordid><originalsourceid>FETCH-LOGICAL-c438t-ca93ceeba58ee6d467ee51df4fc0af1ba85d2f9b1a9530e842e875edd0bea04e3</originalsourceid><addsrcrecordid>eNp1kc1u1TAQhS0EolXpnhXyDhZNaydxYi9R-WmlSmzo2prY44tLYqe2A3THjhfgCXkScnvbigVdzejoOzOaM4S85OyYMylPaib6iqu-OwHTgLVPyP6D9PSffo8c5nzFGONd3XVKPCd7TdtLJaTcJ7_eIc404JJgpFO0OGbqYqKwlDhBQUunZSy-KpC_UuthE2Iu3tBsINAJAmxwwlD-_Px9vcDoyw3F8AWCuVWP6DeP36kZIWfvvIHiY6AQLE04x1ToBgOmW_UFeeZgzHh4Vw_I5Yf3n0_PqotPH89P315Upm1kqQyoxiAOICRiZ9uuRxTcutYZBo4PIIWtnRo4KNEwlG2NshdoLRsQWIvNAXmzmzuneL1gLnry2eA4QsC4ZF0LpVgvatWtKNuhJsWcEzo9Jz9ButGc6e0H9DZivY1Y7z6wWl7dTV-GCe2D4T7vFXi9A3yc9VVcUliP1cOMP7TUXDMuGOd6tm4lj_5DPrr5Lxd1o7k</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2599075296</pqid></control><display><type>article</type><title>Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation</title><source>MEDLINE</source><source>IOP Publishing Journals</source><source>Institute of Physics (IOP) Journals - HEAL-Link</source><creator>K, Karthik ; Kamath S, Sowmya</creator><creatorcontrib>K, Karthik ; Kamath S, Sowmya</creatorcontrib><description>The detailed physiological perspectives captured by medical imaging provides actionable insights to doctors to manage comprehensive care of patients. However, the quality of such diagnostic image modalities is often affected by mismanagement of the image capturing process by poorly trained technicians and older/poorly maintained imaging equipment. Further, a patient is often subjected to scanning at different orientations to capture the frontal, lateral and sagittal views of the affected areas. Due to the large volume of diagnostic scans performed at a modern hospital, adequate documentation of such additional perspectives is mostly overlooked, which is also an essential key element of quality diagnostic systems and predictive analytics systems. Another crucial challenge affecting effective medical image data management is that the diagnostic scans are essentially stored as unstructured data, lacking a well-defined processing methodology for enabling intelligent image data management for supporting applications like similar patient retrieval , automated disease prediction etc. One solution is to incorporate automated diagnostic image descriptions of the observation/findings by leveraging computer vision and natural language processing. In this work, we present multi-task neural models capable of addressing these critical challenges. We propose ESRGAN, an image enhancement technique for improving the quality and visualization of medical chest x-ray images, thereby substantially improving the potential for accurate diagnosis, automatic detection and region-of-interest segmentation. We also propose a CNN-based model called ViewNet for predicting the view orientation of the x-ray image and generating a medical report using Xception net, thus facilitating a robust medical image management system for intelligent diagnosis applications. Experimental results are demonstrated using standard metrics like BRISQUE, PIQE and BLEU scores, indicating that the proposed models achieved excellent performance. Further, the proposed deep learning approaches enable diagnosis in a lesser time and their hybrid architecture shows significant potential for supporting many intelligent diagnosis applications.</description><identifier>ISSN: 2057-1976</identifier><identifier>EISSN: 2057-1976</identifier><identifier>DOI: 10.1088/2057-1976/ac3add</identifier><identifier>PMID: 34789588</identifier><language>eng</language><publisher>England: IOP Publishing</publisher><subject>deep learning ; enhancement ; ESRGAN ; Humans ; medical report ; natural language processing ; Neural Networks, Computer ; orientation ; Radiography ; Radionuclide Imaging ; ViewNet</subject><ispartof>Biomedical physics &amp; engineering express, 2021-11, Vol.8 (1), p.15011</ispartof><rights>2021 IOP Publishing Ltd</rights><rights>2021 IOP Publishing Ltd.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c438t-ca93ceeba58ee6d467ee51df4fc0af1ba85d2f9b1a9530e842e875edd0bea04e3</citedby><cites>FETCH-LOGICAL-c438t-ca93ceeba58ee6d467ee51df4fc0af1ba85d2f9b1a9530e842e875edd0bea04e3</cites><orcidid>0000-0003-0846-2982 ; 0000-0002-0888-7238</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://iopscience.iop.org/article/10.1088/2057-1976/ac3add/pdf$$EPDF$$P50$$Giop$$H</linktopdf><link.rule.ids>314,776,780,27901,27902,53821,53868</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/34789588$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>K, Karthik</creatorcontrib><creatorcontrib>Kamath S, Sowmya</creatorcontrib><title>Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation</title><title>Biomedical physics &amp; engineering express</title><addtitle>BPEX</addtitle><addtitle>Biomed. Phys. Eng. Express</addtitle><description>The detailed physiological perspectives captured by medical imaging provides actionable insights to doctors to manage comprehensive care of patients. However, the quality of such diagnostic image modalities is often affected by mismanagement of the image capturing process by poorly trained technicians and older/poorly maintained imaging equipment. Further, a patient is often subjected to scanning at different orientations to capture the frontal, lateral and sagittal views of the affected areas. Due to the large volume of diagnostic scans performed at a modern hospital, adequate documentation of such additional perspectives is mostly overlooked, which is also an essential key element of quality diagnostic systems and predictive analytics systems. Another crucial challenge affecting effective medical image data management is that the diagnostic scans are essentially stored as unstructured data, lacking a well-defined processing methodology for enabling intelligent image data management for supporting applications like similar patient retrieval , automated disease prediction etc. One solution is to incorporate automated diagnostic image descriptions of the observation/findings by leveraging computer vision and natural language processing. In this work, we present multi-task neural models capable of addressing these critical challenges. We propose ESRGAN, an image enhancement technique for improving the quality and visualization of medical chest x-ray images, thereby substantially improving the potential for accurate diagnosis, automatic detection and region-of-interest segmentation. We also propose a CNN-based model called ViewNet for predicting the view orientation of the x-ray image and generating a medical report using Xception net, thus facilitating a robust medical image management system for intelligent diagnosis applications. Experimental results are demonstrated using standard metrics like BRISQUE, PIQE and BLEU scores, indicating that the proposed models achieved excellent performance. Further, the proposed deep learning approaches enable diagnosis in a lesser time and their hybrid architecture shows significant potential for supporting many intelligent diagnosis applications.</description><subject>deep learning</subject><subject>enhancement</subject><subject>ESRGAN</subject><subject>Humans</subject><subject>medical report</subject><subject>natural language processing</subject><subject>Neural Networks, Computer</subject><subject>orientation</subject><subject>Radiography</subject><subject>Radionuclide Imaging</subject><subject>ViewNet</subject><issn>2057-1976</issn><issn>2057-1976</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>EIF</sourceid><recordid>eNp1kc1u1TAQhS0EolXpnhXyDhZNaydxYi9R-WmlSmzo2prY44tLYqe2A3THjhfgCXkScnvbigVdzejoOzOaM4S85OyYMylPaib6iqu-OwHTgLVPyP6D9PSffo8c5nzFGONd3XVKPCd7TdtLJaTcJ7_eIc404JJgpFO0OGbqYqKwlDhBQUunZSy-KpC_UuthE2Iu3tBsINAJAmxwwlD-_Px9vcDoyw3F8AWCuVWP6DeP36kZIWfvvIHiY6AQLE04x1ToBgOmW_UFeeZgzHh4Vw_I5Yf3n0_PqotPH89P315Upm1kqQyoxiAOICRiZ9uuRxTcutYZBo4PIIWtnRo4KNEwlG2NshdoLRsQWIvNAXmzmzuneL1gLnry2eA4QsC4ZF0LpVgvatWtKNuhJsWcEzo9Jz9ButGc6e0H9DZivY1Y7z6wWl7dTV-GCe2D4T7vFXi9A3yc9VVcUliP1cOMP7TUXDMuGOd6tm4lj_5DPrr5Lxd1o7k</recordid><startdate>20211130</startdate><enddate>20211130</enddate><creator>K, Karthik</creator><creator>Kamath S, Sowmya</creator><general>IOP Publishing</general><scope>CGR</scope><scope>CUY</scope><scope>CVF</scope><scope>ECM</scope><scope>EIF</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0003-0846-2982</orcidid><orcidid>https://orcid.org/0000-0002-0888-7238</orcidid></search><sort><creationdate>20211130</creationdate><title>Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation</title><author>K, Karthik ; Kamath S, Sowmya</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c438t-ca93ceeba58ee6d467ee51df4fc0af1ba85d2f9b1a9530e842e875edd0bea04e3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>deep learning</topic><topic>enhancement</topic><topic>ESRGAN</topic><topic>Humans</topic><topic>medical report</topic><topic>natural language processing</topic><topic>Neural Networks, Computer</topic><topic>orientation</topic><topic>Radiography</topic><topic>Radionuclide Imaging</topic><topic>ViewNet</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>K, Karthik</creatorcontrib><creatorcontrib>Kamath S, Sowmya</creatorcontrib><collection>Medline</collection><collection>MEDLINE</collection><collection>MEDLINE (Ovid)</collection><collection>MEDLINE</collection><collection>MEDLINE</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>Biomedical physics &amp; engineering express</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>K, Karthik</au><au>Kamath S, Sowmya</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation</atitle><jtitle>Biomedical physics &amp; engineering express</jtitle><stitle>BPEX</stitle><addtitle>Biomed. Phys. Eng. Express</addtitle><date>2021-11-30</date><risdate>2021</risdate><volume>8</volume><issue>1</issue><spage>15011</spage><pages>15011-</pages><issn>2057-1976</issn><eissn>2057-1976</eissn><abstract>The detailed physiological perspectives captured by medical imaging provides actionable insights to doctors to manage comprehensive care of patients. However, the quality of such diagnostic image modalities is often affected by mismanagement of the image capturing process by poorly trained technicians and older/poorly maintained imaging equipment. Further, a patient is often subjected to scanning at different orientations to capture the frontal, lateral and sagittal views of the affected areas. Due to the large volume of diagnostic scans performed at a modern hospital, adequate documentation of such additional perspectives is mostly overlooked, which is also an essential key element of quality diagnostic systems and predictive analytics systems. Another crucial challenge affecting effective medical image data management is that the diagnostic scans are essentially stored as unstructured data, lacking a well-defined processing methodology for enabling intelligent image data management for supporting applications like similar patient retrieval , automated disease prediction etc. One solution is to incorporate automated diagnostic image descriptions of the observation/findings by leveraging computer vision and natural language processing. In this work, we present multi-task neural models capable of addressing these critical challenges. We propose ESRGAN, an image enhancement technique for improving the quality and visualization of medical chest x-ray images, thereby substantially improving the potential for accurate diagnosis, automatic detection and region-of-interest segmentation. We also propose a CNN-based model called ViewNet for predicting the view orientation of the x-ray image and generating a medical report using Xception net, thus facilitating a robust medical image management system for intelligent diagnosis applications. Experimental results are demonstrated using standard metrics like BRISQUE, PIQE and BLEU scores, indicating that the proposed models achieved excellent performance. Further, the proposed deep learning approaches enable diagnosis in a lesser time and their hybrid architecture shows significant potential for supporting many intelligent diagnosis applications.</abstract><cop>England</cop><pub>IOP Publishing</pub><pmid>34789588</pmid><doi>10.1088/2057-1976/ac3add</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0003-0846-2982</orcidid><orcidid>https://orcid.org/0000-0002-0888-7238</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 2057-1976
ispartof Biomedical physics & engineering express, 2021-11, Vol.8 (1), p.15011
issn 2057-1976
2057-1976
language eng
recordid cdi_pubmed_primary_34789588
source MEDLINE; IOP Publishing Journals; Institute of Physics (IOP) Journals - HEAL-Link
subjects deep learning
enhancement
ESRGAN
Humans
medical report
natural language processing
Neural Networks, Computer
orientation
Radiography
Radionuclide Imaging
ViewNet
title Deep neural models for automated multi-task diagnostic scan management—quality enhancement, view classification and report generation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-14T12%3A27%3A51IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20neural%20models%20for%20automated%20multi-task%20diagnostic%20scan%20management%E2%80%94quality%20enhancement,%20view%20classification%20and%20report%20generation&rft.jtitle=Biomedical%20physics%20&%20engineering%20express&rft.au=K,%20Karthik&rft.date=2021-11-30&rft.volume=8&rft.issue=1&rft.spage=15011&rft.pages=15011-&rft.issn=2057-1976&rft.eissn=2057-1976&rft_id=info:doi/10.1088/2057-1976/ac3add&rft_dat=%3Cproquest_pubme%3E2599075296%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2599075296&rft_id=info:pmid/34789588&rfr_iscdi=true