Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning
An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which sh...
Gespeichert in:
Veröffentlicht in: | IEEE access 2021-01, Vol.9, p.1-1 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE access |
container_volume | 9 |
creator | Hou, Daibing Zhao, Zijian Liu, Yuying Chang, Faliang Hu, Sanyuan |
description | An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared. |
doi_str_mv | 10.1109/ACCESS.2021.3056175 |
format | Article |
fullrecord | <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2487437961</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9343868</ieee_id><doaj_id>oai_doaj_org_article_bb739072aad0492b871a592e96848a55</doaj_id><sourcerecordid>2487437961</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</originalsourceid><addsrcrecordid>eNpNkV9LwzAUxYsoKOon8CXgc2f-NsljKXMOBoL6oE_htr2dHWsz007YtzezQ8xLwuH8zs3lJMkdozPGqH3Ii2L--jrjlLOZoCpjWp0lV5xlNhVKZOf_3pfJ7TBsaDwmSkpfJR_5fvQdjG1FXnDnw0gW2GOIgu9J4wMpPnEYyXv6Agey7GCNA_lugeT1N4YBQgvbCLZ9tFbYYT-SFULo2359k1w0sB3w9nRfJ2-P87fiKV09L5ZFvkorSc2YMkCjTVNCVZpaom64QmugNI2omZCNjZK0FUqmeGZUzSUwW1uoqUaqM3GdLKfY2sPG7ULbQTg4D637FXxYOwhxvS26stTCUs0hwtLy0mgGynK0mZEGlIpZ91PWLvivfdzbbfw-9PH3jkujpdA2Y9ElJlcV_DAEbP6mMuqOjbipEXdsxJ0aidTdRLWI-EdYIYXJjPgBSXWGsw</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2487437961</pqid></control><display><type>article</type><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><source>DOAJ Directory of Open Access Journals</source><source>IEEE Xplore Open Access Journals</source><source>EZB Electronic Journals Library</source><creator>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</creator><creatorcontrib>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</creatorcontrib><description>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2021.3056175</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Accuracy ; adversarial training ; Chest ; Cider ; Coders ; Decoding ; Diagnostic systems ; Discriminators ; encoder-decoder ; Feature extraction ; Generators ; Learning ; Maximum likelihood estimation ; Medical diagnostic imaging ; Medical imaging ; Medical report generation ; Radiographs ; reinforcement learning ; Report generators ; Semantics ; Training ; X-ray imaging</subject><ispartof>IEEE access, 2021-01, Vol.9, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</citedby><cites>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</cites><orcidid>0000-0002-7849-814X ; 0000-0002-1902-3742 ; 0000-0002-4682-2187 ; 0000-0003-1276-2267</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9343868$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2100,27624,27915,27916,54924</link.rule.ids></links><search><creatorcontrib>Hou, Daibing</creatorcontrib><creatorcontrib>Zhao, Zijian</creatorcontrib><creatorcontrib>Liu, Yuying</creatorcontrib><creatorcontrib>Chang, Faliang</creatorcontrib><creatorcontrib>Hu, Sanyuan</creatorcontrib><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><title>IEEE access</title><addtitle>Access</addtitle><description>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</description><subject>Accuracy</subject><subject>adversarial training</subject><subject>Chest</subject><subject>Cider</subject><subject>Coders</subject><subject>Decoding</subject><subject>Diagnostic systems</subject><subject>Discriminators</subject><subject>encoder-decoder</subject><subject>Feature extraction</subject><subject>Generators</subject><subject>Learning</subject><subject>Maximum likelihood estimation</subject><subject>Medical diagnostic imaging</subject><subject>Medical imaging</subject><subject>Medical report generation</subject><subject>Radiographs</subject><subject>reinforcement learning</subject><subject>Report generators</subject><subject>Semantics</subject><subject>Training</subject><subject>X-ray imaging</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkV9LwzAUxYsoKOon8CXgc2f-NsljKXMOBoL6oE_htr2dHWsz007YtzezQ8xLwuH8zs3lJMkdozPGqH3Ii2L--jrjlLOZoCpjWp0lV5xlNhVKZOf_3pfJ7TBsaDwmSkpfJR_5fvQdjG1FXnDnw0gW2GOIgu9J4wMpPnEYyXv6Agey7GCNA_lugeT1N4YBQgvbCLZ9tFbYYT-SFULo2359k1w0sB3w9nRfJ2-P87fiKV09L5ZFvkorSc2YMkCjTVNCVZpaom64QmugNI2omZCNjZK0FUqmeGZUzSUwW1uoqUaqM3GdLKfY2sPG7ULbQTg4D637FXxYOwhxvS26stTCUs0hwtLy0mgGynK0mZEGlIpZ91PWLvivfdzbbfw-9PH3jkujpdA2Y9ElJlcV_DAEbP6mMuqOjbipEXdsxJ0aidTdRLWI-EdYIYXJjPgBSXWGsw</recordid><startdate>20210101</startdate><enddate>20210101</enddate><creator>Hou, Daibing</creator><creator>Zhao, Zijian</creator><creator>Liu, Yuying</creator><creator>Chang, Faliang</creator><creator>Hu, Sanyuan</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7849-814X</orcidid><orcidid>https://orcid.org/0000-0002-1902-3742</orcidid><orcidid>https://orcid.org/0000-0002-4682-2187</orcidid><orcidid>https://orcid.org/0000-0003-1276-2267</orcidid></search><sort><creationdate>20210101</creationdate><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><author>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Accuracy</topic><topic>adversarial training</topic><topic>Chest</topic><topic>Cider</topic><topic>Coders</topic><topic>Decoding</topic><topic>Diagnostic systems</topic><topic>Discriminators</topic><topic>encoder-decoder</topic><topic>Feature extraction</topic><topic>Generators</topic><topic>Learning</topic><topic>Maximum likelihood estimation</topic><topic>Medical diagnostic imaging</topic><topic>Medical imaging</topic><topic>Medical report generation</topic><topic>Radiographs</topic><topic>reinforcement learning</topic><topic>Report generators</topic><topic>Semantics</topic><topic>Training</topic><topic>X-ray imaging</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hou, Daibing</creatorcontrib><creatorcontrib>Zhao, Zijian</creatorcontrib><creatorcontrib>Liu, Yuying</creatorcontrib><creatorcontrib>Chang, Faliang</creatorcontrib><creatorcontrib>Hu, Sanyuan</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics & Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hou, Daibing</au><au>Zhao, Zijian</au><au>Liu, Yuying</au><au>Chang, Faliang</au><au>Hu, Sanyuan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2021-01-01</date><risdate>2021</risdate><volume>9</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2021.3056175</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0002-7849-814X</orcidid><orcidid>https://orcid.org/0000-0002-1902-3742</orcidid><orcidid>https://orcid.org/0000-0002-4682-2187</orcidid><orcidid>https://orcid.org/0000-0003-1276-2267</orcidid><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2169-3536 |
ispartof | IEEE access, 2021-01, Vol.9, p.1-1 |
issn | 2169-3536 2169-3536 |
language | eng |
recordid | cdi_proquest_journals_2487437961 |
source | DOAJ Directory of Open Access Journals; IEEE Xplore Open Access Journals; EZB Electronic Journals Library |
subjects | Accuracy adversarial training Chest Cider Coders Decoding Diagnostic systems Discriminators encoder-decoder Feature extraction Generators Learning Maximum likelihood estimation Medical diagnostic imaging Medical imaging Medical report generation Radiographs reinforcement learning Report generators Semantics Training X-ray imaging |
title | Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T19%3A52%3A39IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Automatic%20Report%20Generation%20for%20Chest%20X-Ray%20Images%20via%20Adversarial%20Reinforcement%20Learning&rft.jtitle=IEEE%20access&rft.au=Hou,%20Daibing&rft.date=2021-01-01&rft.volume=9&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2021.3056175&rft_dat=%3Cproquest_ieee_%3E2487437961%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2487437961&rft_id=info:pmid/&rft_ieee_id=9343868&rft_doaj_id=oai_doaj_org_article_bb739072aad0492b871a592e96848a55&rfr_iscdi=true |