Deep learning attention-guided radiomics for COVID-19 chest radiograph classification

Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differen...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Quantitative imaging in medicine and surgery 2023-02, Vol.13 (2), p.572-584
Hauptverfasser: Yang, Dongrong, Ren, Ge, Ni, Ruiyan, Huang, Yu-Hua, Lam, Ngo Fung Daniel, Sun, Hongfei, Wan, Shiu Bun Nelson, Wong, Man Fung Esther, Chan, King Kwong, Tsang, Hoi Ching Hailey, Xu, Lu, Wu, Tak Chiu, Kong, Feng-Ming Spring, Wáng, Yì Xiáng J, Qin, Jing, Chan, Lawrence Wing Chi, Ying, Michael, Cai, Jing
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 584
container_issue 2
container_start_page 572
container_title Quantitative imaging in medicine and surgery
container_volume 13
creator Yang, Dongrong
Ren, Ge
Ni, Ruiyan
Huang, Yu-Hua
Lam, Ngo Fung Daniel
Sun, Hongfei
Wan, Shiu Bun Nelson
Wong, Man Fung Esther
Chan, King Kwong
Tsang, Hoi Ching Hailey
Xu, Lu
Wu, Tak Chiu
Kong, Feng-Ming Spring
Wáng, Yì Xiáng J
Qin, Jing
Chan, Lawrence Wing Chi
Ying, Michael
Cai, Jing
description Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value
doi_str_mv 10.21037/qims-22-531
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9929417</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2779349619</sourcerecordid><originalsourceid>FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</originalsourceid><addsrcrecordid>eNpVUclOwzAQtRCIVqU3zihHDgTscRzHFyTUslSq1AvlajmOkxplq50g8fekdBHMZUZ6b94sD6Frgu-BYMoftrbyIUDIKDlDYwCgYURxfH6sQcAITb3_xEPwhHCCL9GIxgkREIsxWs-NaYPSKFfbughU15m6s00dFr3NTBY4ldmmstoHeeOC2epjMQ-JCPTG-G4PFk61m0CXynubW6123VfoIlelN9NDnqD1y_P77C1crl4Xs6dlqGkSdaHSwHOWqERlKWCVZYzmaZzmigmaAMS5YNRgwwhwwgQjDOuhwBREFnMBCZ2gx71u26eVyfSwu1OlbJ2tlPuWjbLyP1LbjSyaLykEiIjwQeD2IOCabT_cJCvrtSlLVZum9xI4FzQSMRED9W5P1a7x3pn8NIZg-WuG3JkhAeRgxkC_-bvaiXx8Pf0B0CiGAw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2779349619</pqid></control><display><type>article</type><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><source>PubMed Central</source><creator>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</creator><creatorcontrib>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</creatorcontrib><description>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value &lt;0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19). A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</description><identifier>ISSN: 2223-4292</identifier><identifier>EISSN: 2223-4306</identifier><identifier>DOI: 10.21037/qims-22-531</identifier><identifier>PMID: 36819269</identifier><language>eng</language><publisher>China: AME Publishing Company</publisher><subject>Original</subject><ispartof>Quantitative imaging in medicine and surgery, 2023-02, Vol.13 (2), p.572-584</ispartof><rights>2023 Quantitative Imaging in Medicine and Surgery. All rights reserved.</rights><rights>2023 Quantitative Imaging in Medicine and Surgery. All rights reserved. 2023 Quantitative Imaging in Medicine and Surgery.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/pdf/$$EPDF$$P50$$Gpubmedcentral$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/$$EHTML$$P50$$Gpubmedcentral$$Hfree_for_read</linktohtml><link.rule.ids>230,315,728,781,785,886,27929,27930,53796,53798</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36819269$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Yang, Dongrong</creatorcontrib><creatorcontrib>Ren, Ge</creatorcontrib><creatorcontrib>Ni, Ruiyan</creatorcontrib><creatorcontrib>Huang, Yu-Hua</creatorcontrib><creatorcontrib>Lam, Ngo Fung Daniel</creatorcontrib><creatorcontrib>Sun, Hongfei</creatorcontrib><creatorcontrib>Wan, Shiu Bun Nelson</creatorcontrib><creatorcontrib>Wong, Man Fung Esther</creatorcontrib><creatorcontrib>Chan, King Kwong</creatorcontrib><creatorcontrib>Tsang, Hoi Ching Hailey</creatorcontrib><creatorcontrib>Xu, Lu</creatorcontrib><creatorcontrib>Wu, Tak Chiu</creatorcontrib><creatorcontrib>Kong, Feng-Ming Spring</creatorcontrib><creatorcontrib>Wáng, Yì Xiáng J</creatorcontrib><creatorcontrib>Qin, Jing</creatorcontrib><creatorcontrib>Chan, Lawrence Wing Chi</creatorcontrib><creatorcontrib>Ying, Michael</creatorcontrib><creatorcontrib>Cai, Jing</creatorcontrib><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><title>Quantitative imaging in medicine and surgery</title><addtitle>Quant Imaging Med Surg</addtitle><description>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value &lt;0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19). A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</description><subject>Original</subject><issn>2223-4292</issn><issn>2223-4306</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpVUclOwzAQtRCIVqU3zihHDgTscRzHFyTUslSq1AvlajmOkxplq50g8fekdBHMZUZ6b94sD6Frgu-BYMoftrbyIUDIKDlDYwCgYURxfH6sQcAITb3_xEPwhHCCL9GIxgkREIsxWs-NaYPSKFfbughU15m6s00dFr3NTBY4ldmmstoHeeOC2epjMQ-JCPTG-G4PFk61m0CXynubW6123VfoIlelN9NDnqD1y_P77C1crl4Xs6dlqGkSdaHSwHOWqERlKWCVZYzmaZzmigmaAMS5YNRgwwhwwgQjDOuhwBREFnMBCZ2gx71u26eVyfSwu1OlbJ2tlPuWjbLyP1LbjSyaLykEiIjwQeD2IOCabT_cJCvrtSlLVZum9xI4FzQSMRED9W5P1a7x3pn8NIZg-WuG3JkhAeRgxkC_-bvaiXx8Pf0B0CiGAw</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Yang, Dongrong</creator><creator>Ren, Ge</creator><creator>Ni, Ruiyan</creator><creator>Huang, Yu-Hua</creator><creator>Lam, Ngo Fung Daniel</creator><creator>Sun, Hongfei</creator><creator>Wan, Shiu Bun Nelson</creator><creator>Wong, Man Fung Esther</creator><creator>Chan, King Kwong</creator><creator>Tsang, Hoi Ching Hailey</creator><creator>Xu, Lu</creator><creator>Wu, Tak Chiu</creator><creator>Kong, Feng-Ming Spring</creator><creator>Wáng, Yì Xiáng J</creator><creator>Qin, Jing</creator><creator>Chan, Lawrence Wing Chi</creator><creator>Ying, Michael</creator><creator>Cai, Jing</creator><general>AME Publishing Company</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope></search><sort><creationdate>20230201</creationdate><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><author>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Original</topic><toplevel>online_resources</toplevel><creatorcontrib>Yang, Dongrong</creatorcontrib><creatorcontrib>Ren, Ge</creatorcontrib><creatorcontrib>Ni, Ruiyan</creatorcontrib><creatorcontrib>Huang, Yu-Hua</creatorcontrib><creatorcontrib>Lam, Ngo Fung Daniel</creatorcontrib><creatorcontrib>Sun, Hongfei</creatorcontrib><creatorcontrib>Wan, Shiu Bun Nelson</creatorcontrib><creatorcontrib>Wong, Man Fung Esther</creatorcontrib><creatorcontrib>Chan, King Kwong</creatorcontrib><creatorcontrib>Tsang, Hoi Ching Hailey</creatorcontrib><creatorcontrib>Xu, Lu</creatorcontrib><creatorcontrib>Wu, Tak Chiu</creatorcontrib><creatorcontrib>Kong, Feng-Ming Spring</creatorcontrib><creatorcontrib>Wáng, Yì Xiáng J</creatorcontrib><creatorcontrib>Qin, Jing</creatorcontrib><creatorcontrib>Chan, Lawrence Wing Chi</creatorcontrib><creatorcontrib>Ying, Michael</creatorcontrib><creatorcontrib>Cai, Jing</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Quantitative imaging in medicine and surgery</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, Dongrong</au><au>Ren, Ge</au><au>Ni, Ruiyan</au><au>Huang, Yu-Hua</au><au>Lam, Ngo Fung Daniel</au><au>Sun, Hongfei</au><au>Wan, Shiu Bun Nelson</au><au>Wong, Man Fung Esther</au><au>Chan, King Kwong</au><au>Tsang, Hoi Ching Hailey</au><au>Xu, Lu</au><au>Wu, Tak Chiu</au><au>Kong, Feng-Ming Spring</au><au>Wáng, Yì Xiáng J</au><au>Qin, Jing</au><au>Chan, Lawrence Wing Chi</au><au>Ying, Michael</au><au>Cai, Jing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</atitle><jtitle>Quantitative imaging in medicine and surgery</jtitle><addtitle>Quant Imaging Med Surg</addtitle><date>2023-02-01</date><risdate>2023</risdate><volume>13</volume><issue>2</issue><spage>572</spage><epage>584</epage><pages>572-584</pages><issn>2223-4292</issn><eissn>2223-4306</eissn><abstract>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR). In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation. Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value &lt;0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19). A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</abstract><cop>China</cop><pub>AME Publishing Company</pub><pmid>36819269</pmid><doi>10.21037/qims-22-531</doi><tpages>13</tpages><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2223-4292
ispartof Quantitative imaging in medicine and surgery, 2023-02, Vol.13 (2), p.572-584
issn 2223-4292
2223-4306
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9929417
source Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals; PubMed Central
subjects Original
title Deep learning attention-guided radiomics for COVID-19 chest radiograph classification
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-14T13%3A47%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20learning%20attention-guided%20radiomics%20for%20COVID-19%20chest%20radiograph%20classification&rft.jtitle=Quantitative%20imaging%20in%20medicine%20and%20surgery&rft.au=Yang,%20Dongrong&rft.date=2023-02-01&rft.volume=13&rft.issue=2&rft.spage=572&rft.epage=584&rft.pages=572-584&rft.issn=2223-4292&rft.eissn=2223-4306&rft_id=info:doi/10.21037/qims-22-531&rft_dat=%3Cproquest_pubme%3E2779349619%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2779349619&rft_id=info:pmid/36819269&rfr_iscdi=true