Deep learning attention-guided radiomics for COVID-19 chest radiograph classification
Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differen...
Gespeichert in:
Veröffentlicht in: | Quantitative imaging in medicine and surgery 2023-02, Vol.13 (2), p.572-584 |
---|---|
Hauptverfasser: | , , , , , , , , , , , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 584 |
---|---|
container_issue | 2 |
container_start_page | 572 |
container_title | Quantitative imaging in medicine and surgery |
container_volume | 13 |
creator | Yang, Dongrong Ren, Ge Ni, Ruiyan Huang, Yu-Hua Lam, Ngo Fung Daniel Sun, Hongfei Wan, Shiu Bun Nelson Wong, Man Fung Esther Chan, King Kwong Tsang, Hoi Ching Hailey Xu, Lu Wu, Tak Chiu Kong, Feng-Ming Spring Wáng, Yì Xiáng J Qin, Jing Chan, Lawrence Wing Chi Ying, Michael Cai, Jing |
description | Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR).
In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation.
Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value |
doi_str_mv | 10.21037/qims-22-531 |
format | Article |
fullrecord | <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9929417</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2779349619</sourcerecordid><originalsourceid>FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</originalsourceid><addsrcrecordid>eNpVUclOwzAQtRCIVqU3zihHDgTscRzHFyTUslSq1AvlajmOkxplq50g8fekdBHMZUZ6b94sD6Frgu-BYMoftrbyIUDIKDlDYwCgYURxfH6sQcAITb3_xEPwhHCCL9GIxgkREIsxWs-NaYPSKFfbughU15m6s00dFr3NTBY4ldmmstoHeeOC2epjMQ-JCPTG-G4PFk61m0CXynubW6123VfoIlelN9NDnqD1y_P77C1crl4Xs6dlqGkSdaHSwHOWqERlKWCVZYzmaZzmigmaAMS5YNRgwwhwwgQjDOuhwBREFnMBCZ2gx71u26eVyfSwu1OlbJ2tlPuWjbLyP1LbjSyaLykEiIjwQeD2IOCabT_cJCvrtSlLVZum9xI4FzQSMRED9W5P1a7x3pn8NIZg-WuG3JkhAeRgxkC_-bvaiXx8Pf0B0CiGAw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2779349619</pqid></control><display><type>article</type><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><source>PubMed Central</source><creator>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</creator><creatorcontrib>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</creatorcontrib><description>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR).
In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation.
Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value <0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19).
A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</description><identifier>ISSN: 2223-4292</identifier><identifier>EISSN: 2223-4306</identifier><identifier>DOI: 10.21037/qims-22-531</identifier><identifier>PMID: 36819269</identifier><language>eng</language><publisher>China: AME Publishing Company</publisher><subject>Original</subject><ispartof>Quantitative imaging in medicine and surgery, 2023-02, Vol.13 (2), p.572-584</ispartof><rights>2023 Quantitative Imaging in Medicine and Surgery. All rights reserved.</rights><rights>2023 Quantitative Imaging in Medicine and Surgery. All rights reserved. 2023 Quantitative Imaging in Medicine and Surgery.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</citedby></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/pdf/$$EPDF$$P50$$Gpubmedcentral$$Hfree_for_read</linktopdf><linktohtml>$$Uhttps://www.ncbi.nlm.nih.gov/pmc/articles/PMC9929417/$$EHTML$$P50$$Gpubmedcentral$$Hfree_for_read</linktohtml><link.rule.ids>230,315,728,781,785,886,27929,27930,53796,53798</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/36819269$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Yang, Dongrong</creatorcontrib><creatorcontrib>Ren, Ge</creatorcontrib><creatorcontrib>Ni, Ruiyan</creatorcontrib><creatorcontrib>Huang, Yu-Hua</creatorcontrib><creatorcontrib>Lam, Ngo Fung Daniel</creatorcontrib><creatorcontrib>Sun, Hongfei</creatorcontrib><creatorcontrib>Wan, Shiu Bun Nelson</creatorcontrib><creatorcontrib>Wong, Man Fung Esther</creatorcontrib><creatorcontrib>Chan, King Kwong</creatorcontrib><creatorcontrib>Tsang, Hoi Ching Hailey</creatorcontrib><creatorcontrib>Xu, Lu</creatorcontrib><creatorcontrib>Wu, Tak Chiu</creatorcontrib><creatorcontrib>Kong, Feng-Ming Spring</creatorcontrib><creatorcontrib>Wáng, Yì Xiáng J</creatorcontrib><creatorcontrib>Qin, Jing</creatorcontrib><creatorcontrib>Chan, Lawrence Wing Chi</creatorcontrib><creatorcontrib>Ying, Michael</creatorcontrib><creatorcontrib>Cai, Jing</creatorcontrib><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><title>Quantitative imaging in medicine and surgery</title><addtitle>Quant Imaging Med Surg</addtitle><description>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR).
In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation.
Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value <0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19).
A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</description><subject>Original</subject><issn>2223-4292</issn><issn>2223-4306</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpVUclOwzAQtRCIVqU3zihHDgTscRzHFyTUslSq1AvlajmOkxplq50g8fekdBHMZUZ6b94sD6Frgu-BYMoftrbyIUDIKDlDYwCgYURxfH6sQcAITb3_xEPwhHCCL9GIxgkREIsxWs-NaYPSKFfbughU15m6s00dFr3NTBY4ldmmstoHeeOC2epjMQ-JCPTG-G4PFk61m0CXynubW6123VfoIlelN9NDnqD1y_P77C1crl4Xs6dlqGkSdaHSwHOWqERlKWCVZYzmaZzmigmaAMS5YNRgwwhwwgQjDOuhwBREFnMBCZ2gx71u26eVyfSwu1OlbJ2tlPuWjbLyP1LbjSyaLykEiIjwQeD2IOCabT_cJCvrtSlLVZum9xI4FzQSMRED9W5P1a7x3pn8NIZg-WuG3JkhAeRgxkC_-bvaiXx8Pf0B0CiGAw</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Yang, Dongrong</creator><creator>Ren, Ge</creator><creator>Ni, Ruiyan</creator><creator>Huang, Yu-Hua</creator><creator>Lam, Ngo Fung Daniel</creator><creator>Sun, Hongfei</creator><creator>Wan, Shiu Bun Nelson</creator><creator>Wong, Man Fung Esther</creator><creator>Chan, King Kwong</creator><creator>Tsang, Hoi Ching Hailey</creator><creator>Xu, Lu</creator><creator>Wu, Tak Chiu</creator><creator>Kong, Feng-Ming Spring</creator><creator>Wáng, Yì Xiáng J</creator><creator>Qin, Jing</creator><creator>Chan, Lawrence Wing Chi</creator><creator>Ying, Michael</creator><creator>Cai, Jing</creator><general>AME Publishing Company</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><scope>5PM</scope></search><sort><creationdate>20230201</creationdate><title>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</title><author>Yang, Dongrong ; Ren, Ge ; Ni, Ruiyan ; Huang, Yu-Hua ; Lam, Ngo Fung Daniel ; Sun, Hongfei ; Wan, Shiu Bun Nelson ; Wong, Man Fung Esther ; Chan, King Kwong ; Tsang, Hoi Ching Hailey ; Xu, Lu ; Wu, Tak Chiu ; Kong, Feng-Ming Spring ; Wáng, Yì Xiáng J ; Qin, Jing ; Chan, Lawrence Wing Chi ; Ying, Michael ; Cai, Jing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c384t-ac27f58a8adb20add53fb6bfa5938226f953e0e51271595150c7150329d679283</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Original</topic><toplevel>online_resources</toplevel><creatorcontrib>Yang, Dongrong</creatorcontrib><creatorcontrib>Ren, Ge</creatorcontrib><creatorcontrib>Ni, Ruiyan</creatorcontrib><creatorcontrib>Huang, Yu-Hua</creatorcontrib><creatorcontrib>Lam, Ngo Fung Daniel</creatorcontrib><creatorcontrib>Sun, Hongfei</creatorcontrib><creatorcontrib>Wan, Shiu Bun Nelson</creatorcontrib><creatorcontrib>Wong, Man Fung Esther</creatorcontrib><creatorcontrib>Chan, King Kwong</creatorcontrib><creatorcontrib>Tsang, Hoi Ching Hailey</creatorcontrib><creatorcontrib>Xu, Lu</creatorcontrib><creatorcontrib>Wu, Tak Chiu</creatorcontrib><creatorcontrib>Kong, Feng-Ming Spring</creatorcontrib><creatorcontrib>Wáng, Yì Xiáng J</creatorcontrib><creatorcontrib>Qin, Jing</creatorcontrib><creatorcontrib>Chan, Lawrence Wing Chi</creatorcontrib><creatorcontrib>Ying, Michael</creatorcontrib><creatorcontrib>Cai, Jing</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>Quantitative imaging in medicine and surgery</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Yang, Dongrong</au><au>Ren, Ge</au><au>Ni, Ruiyan</au><au>Huang, Yu-Hua</au><au>Lam, Ngo Fung Daniel</au><au>Sun, Hongfei</au><au>Wan, Shiu Bun Nelson</au><au>Wong, Man Fung Esther</au><au>Chan, King Kwong</au><au>Tsang, Hoi Ching Hailey</au><au>Xu, Lu</au><au>Wu, Tak Chiu</au><au>Kong, Feng-Ming Spring</au><au>Wáng, Yì Xiáng J</au><au>Qin, Jing</au><au>Chan, Lawrence Wing Chi</au><au>Ying, Michael</au><au>Cai, Jing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Deep learning attention-guided radiomics for COVID-19 chest radiograph classification</atitle><jtitle>Quantitative imaging in medicine and surgery</jtitle><addtitle>Quant Imaging Med Surg</addtitle><date>2023-02-01</date><risdate>2023</risdate><volume>13</volume><issue>2</issue><spage>572</spage><epage>584</epage><pages>572-584</pages><issn>2223-4292</issn><eissn>2223-4306</eissn><abstract>Accurate assessment of coronavirus disease 2019 (COVID-19) lung involvement through chest radiograph plays an important role in effective management of the infection. This study aims to develop a two-step feature merging method to integrate image features from deep learning and radiomics to differentiate COVID-19, non-COVID-19 pneumonia and normal chest radiographs (CXR).
In this study, a deformable convolutional neural network (deformable CNN) was developed and used as a feature extractor to obtain 1,024-dimensional deep learning latent representation (DLR) features. Then 1,069-dimensional radiomics features were extracted from the region of interest (ROI) guided by deformable CNN's attention. The two feature sets were concatenated to generate a merged feature set for classification. For comparative experiments, the same process has been applied to the DLR-only feature set for verifying the effectiveness of feature concatenation.
Using the merged feature set resulted in an overall average accuracy of 91.0% for three-class classification, representing a statistically significant improvement of 0.6% compared to the DLR-only classification. The recall and precision of classification into the COVID-19 class were 0.926 and 0.976, respectively. The feature merging method was shown to significantly improve the classification performance as compared to using only deep learning features, regardless of choice of classifier (P value <0.0001). Three classes' F1-score were 0.892, 0.890, and 0.950 correspondingly (i.e., normal, non-COVID-19 pneumonia, COVID-19).
A two-step COVID-19 classification framework integrating information from both DLR and radiomics features (guided by deep learning attention mechanism) has been developed. The proposed feature merging method has been shown to improve the performance of chest radiograph classification as compared to the case of using only deep learning features.</abstract><cop>China</cop><pub>AME Publishing Company</pub><pmid>36819269</pmid><doi>10.21037/qims-22-531</doi><tpages>13</tpages><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2223-4292 |
ispartof | Quantitative imaging in medicine and surgery, 2023-02, Vol.13 (2), p.572-584 |
issn | 2223-4292 2223-4306 |
language | eng |
recordid | cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_9929417 |
source | Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals; PubMed Central |
subjects | Original |
title | Deep learning attention-guided radiomics for COVID-19 chest radiograph classification |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-14T13%3A47%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Deep%20learning%20attention-guided%20radiomics%20for%20COVID-19%20chest%20radiograph%20classification&rft.jtitle=Quantitative%20imaging%20in%20medicine%20and%20surgery&rft.au=Yang,%20Dongrong&rft.date=2023-02-01&rft.volume=13&rft.issue=2&rft.spage=572&rft.epage=584&rft.pages=572-584&rft.issn=2223-4292&rft.eissn=2223-4306&rft_id=info:doi/10.21037/qims-22-531&rft_dat=%3Cproquest_pubme%3E2779349619%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2779349619&rft_id=info:pmid/36819269&rfr_iscdi=true |