Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning

An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which sh...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2021-01, Vol.9, p.1-1
Hauptverfasser: Hou, Daibing, Zhao, Zijian, Liu, Yuying, Chang, Faliang, Hu, Sanyuan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue
container_start_page 1
container_title IEEE access
container_volume 9
creator Hou, Daibing
Zhao, Zijian
Liu, Yuying
Chang, Faliang
Hu, Sanyuan
description An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.
doi_str_mv 10.1109/ACCESS.2021.3056175
format Article
fullrecord <record><control><sourceid>proquest_ieee_</sourceid><recordid>TN_cdi_proquest_journals_2487437961</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>9343868</ieee_id><doaj_id>oai_doaj_org_article_bb739072aad0492b871a592e96848a55</doaj_id><sourcerecordid>2487437961</sourcerecordid><originalsourceid>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</originalsourceid><addsrcrecordid>eNpNkV9LwzAUxYsoKOon8CXgc2f-NsljKXMOBoL6oE_htr2dHWsz007YtzezQ8xLwuH8zs3lJMkdozPGqH3Ii2L--jrjlLOZoCpjWp0lV5xlNhVKZOf_3pfJ7TBsaDwmSkpfJR_5fvQdjG1FXnDnw0gW2GOIgu9J4wMpPnEYyXv6Agey7GCNA_lugeT1N4YBQgvbCLZ9tFbYYT-SFULo2359k1w0sB3w9nRfJ2-P87fiKV09L5ZFvkorSc2YMkCjTVNCVZpaom64QmugNI2omZCNjZK0FUqmeGZUzSUwW1uoqUaqM3GdLKfY2sPG7ULbQTg4D637FXxYOwhxvS26stTCUs0hwtLy0mgGynK0mZEGlIpZ91PWLvivfdzbbfw-9PH3jkujpdA2Y9ElJlcV_DAEbP6mMuqOjbipEXdsxJ0aidTdRLWI-EdYIYXJjPgBSXWGsw</addsrcrecordid><sourcetype>Open Website</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2487437961</pqid></control><display><type>article</type><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><source>DOAJ Directory of Open Access Journals</source><source>IEEE Xplore Open Access Journals</source><source>EZB Electronic Journals Library</source><creator>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</creator><creatorcontrib>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</creatorcontrib><description>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</description><identifier>ISSN: 2169-3536</identifier><identifier>EISSN: 2169-3536</identifier><identifier>DOI: 10.1109/ACCESS.2021.3056175</identifier><identifier>CODEN: IAECCG</identifier><language>eng</language><publisher>Piscataway: IEEE</publisher><subject>Accuracy ; adversarial training ; Chest ; Cider ; Coders ; Decoding ; Diagnostic systems ; Discriminators ; encoder-decoder ; Feature extraction ; Generators ; Learning ; Maximum likelihood estimation ; Medical diagnostic imaging ; Medical imaging ; Medical report generation ; Radiographs ; reinforcement learning ; Report generators ; Semantics ; Training ; X-ray imaging</subject><ispartof>IEEE access, 2021-01, Vol.9, p.1-1</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2021</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</citedby><cites>FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</cites><orcidid>0000-0002-7849-814X ; 0000-0002-1902-3742 ; 0000-0002-4682-2187 ; 0000-0003-1276-2267</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/9343868$$EHTML$$P50$$Gieee$$Hfree_for_read</linktohtml><link.rule.ids>314,780,784,864,2100,27624,27915,27916,54924</link.rule.ids></links><search><creatorcontrib>Hou, Daibing</creatorcontrib><creatorcontrib>Zhao, Zijian</creatorcontrib><creatorcontrib>Liu, Yuying</creatorcontrib><creatorcontrib>Chang, Faliang</creatorcontrib><creatorcontrib>Hu, Sanyuan</creatorcontrib><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><title>IEEE access</title><addtitle>Access</addtitle><description>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</description><subject>Accuracy</subject><subject>adversarial training</subject><subject>Chest</subject><subject>Cider</subject><subject>Coders</subject><subject>Decoding</subject><subject>Diagnostic systems</subject><subject>Discriminators</subject><subject>encoder-decoder</subject><subject>Feature extraction</subject><subject>Generators</subject><subject>Learning</subject><subject>Maximum likelihood estimation</subject><subject>Medical diagnostic imaging</subject><subject>Medical imaging</subject><subject>Medical report generation</subject><subject>Radiographs</subject><subject>reinforcement learning</subject><subject>Report generators</subject><subject>Semantics</subject><subject>Training</subject><subject>X-ray imaging</subject><issn>2169-3536</issn><issn>2169-3536</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>ESBDL</sourceid><sourceid>RIE</sourceid><sourceid>DOA</sourceid><recordid>eNpNkV9LwzAUxYsoKOon8CXgc2f-NsljKXMOBoL6oE_htr2dHWsz007YtzezQ8xLwuH8zs3lJMkdozPGqH3Ii2L--jrjlLOZoCpjWp0lV5xlNhVKZOf_3pfJ7TBsaDwmSkpfJR_5fvQdjG1FXnDnw0gW2GOIgu9J4wMpPnEYyXv6Agey7GCNA_lugeT1N4YBQgvbCLZ9tFbYYT-SFULo2359k1w0sB3w9nRfJ2-P87fiKV09L5ZFvkorSc2YMkCjTVNCVZpaom64QmugNI2omZCNjZK0FUqmeGZUzSUwW1uoqUaqM3GdLKfY2sPG7ULbQTg4D637FXxYOwhxvS26stTCUs0hwtLy0mgGynK0mZEGlIpZ91PWLvivfdzbbfw-9PH3jkujpdA2Y9ElJlcV_DAEbP6mMuqOjbipEXdsxJ0aidTdRLWI-EdYIYXJjPgBSXWGsw</recordid><startdate>20210101</startdate><enddate>20210101</enddate><creator>Hou, Daibing</creator><creator>Zhao, Zijian</creator><creator>Liu, Yuying</creator><creator>Chang, Faliang</creator><creator>Hu, Sanyuan</creator><general>IEEE</general><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>97E</scope><scope>ESBDL</scope><scope>RIA</scope><scope>RIE</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>7SP</scope><scope>7SR</scope><scope>8BQ</scope><scope>8FD</scope><scope>JG9</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>DOA</scope><orcidid>https://orcid.org/0000-0002-7849-814X</orcidid><orcidid>https://orcid.org/0000-0002-1902-3742</orcidid><orcidid>https://orcid.org/0000-0002-4682-2187</orcidid><orcidid>https://orcid.org/0000-0003-1276-2267</orcidid></search><sort><creationdate>20210101</creationdate><title>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</title><author>Hou, Daibing ; Zhao, Zijian ; Liu, Yuying ; Chang, Faliang ; Hu, Sanyuan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c408t-1ae878fbacb8d4e7f25e98ab8f3d134f9e7f49ce4152685d24a19d9ad07e0763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Accuracy</topic><topic>adversarial training</topic><topic>Chest</topic><topic>Cider</topic><topic>Coders</topic><topic>Decoding</topic><topic>Diagnostic systems</topic><topic>Discriminators</topic><topic>encoder-decoder</topic><topic>Feature extraction</topic><topic>Generators</topic><topic>Learning</topic><topic>Maximum likelihood estimation</topic><topic>Medical diagnostic imaging</topic><topic>Medical imaging</topic><topic>Medical report generation</topic><topic>Radiographs</topic><topic>reinforcement learning</topic><topic>Report generators</topic><topic>Semantics</topic><topic>Training</topic><topic>X-ray imaging</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Hou, Daibing</creatorcontrib><creatorcontrib>Zhao, Zijian</creatorcontrib><creatorcontrib>Liu, Yuying</creatorcontrib><creatorcontrib>Chang, Faliang</creatorcontrib><creatorcontrib>Hu, Sanyuan</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE Xplore Open Access Journals</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Electronics &amp; Communications Abstracts</collection><collection>Engineered Materials Abstracts</collection><collection>METADEX</collection><collection>Technology Research Database</collection><collection>Materials Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>DOAJ Directory of Open Access Journals</collection><jtitle>IEEE access</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Hou, Daibing</au><au>Zhao, Zijian</au><au>Liu, Yuying</au><au>Chang, Faliang</au><au>Hu, Sanyuan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning</atitle><jtitle>IEEE access</jtitle><stitle>Access</stitle><date>2021-01-01</date><risdate>2021</risdate><volume>9</volume><spage>1</spage><epage>1</epage><pages>1-1</pages><issn>2169-3536</issn><eissn>2169-3536</eissn><coden>IAECCG</coden><abstract>An adversarial reinforced report-generation framework for chest x-ray images is proposed. Previous medical-report-generation models are mostly trained by minimizing the cross-entropy loss or further optimizing the common image-captioning metrics, such as CIDEr, ignoring diagnostic accuracy, which should be the first consideration in this area. Inspired by the generative adversarial network, an adversarial reinforcement learning approach is proposed for report generation of chest x-ray images considering both diagnostic accuracy and language fluency. Specifically, an accuracy discriminator (AD) and fluency discriminator (FD) are built that serve as the evaluators by which a report based on these two aspects is scored. The FD checks how likely a report originates from a human expert, while the AD determines how much a report covers the key chest observations. The weighted score is viewed as a "reward" used for training the report generator via reinforcement learning, which solves the problem that the gradient cannot be passed back to the generative model when the output is discrete. Simultaneously, these two discriminators are optimized by maximum-likelihood estimation for better assessment ability. Additionally, a multi-type medical concept fused encoder followed by a hierarchical decoder is adopted as the report generator. Experiments on two large radiograph datasets demonstrate that the proposed model outperforms all methods to which it is compared.</abstract><cop>Piscataway</cop><pub>IEEE</pub><doi>10.1109/ACCESS.2021.3056175</doi><tpages>1</tpages><orcidid>https://orcid.org/0000-0002-7849-814X</orcidid><orcidid>https://orcid.org/0000-0002-1902-3742</orcidid><orcidid>https://orcid.org/0000-0002-4682-2187</orcidid><orcidid>https://orcid.org/0000-0003-1276-2267</orcidid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2169-3536
ispartof IEEE access, 2021-01, Vol.9, p.1-1
issn 2169-3536
2169-3536
language eng
recordid cdi_proquest_journals_2487437961
source DOAJ Directory of Open Access Journals; IEEE Xplore Open Access Journals; EZB Electronic Journals Library
subjects Accuracy
adversarial training
Chest
Cider
Coders
Decoding
Diagnostic systems
Discriminators
encoder-decoder
Feature extraction
Generators
Learning
Maximum likelihood estimation
Medical diagnostic imaging
Medical imaging
Medical report generation
Radiographs
reinforcement learning
Report generators
Semantics
Training
X-ray imaging
title Automatic Report Generation for Chest X-Ray Images via Adversarial Reinforcement Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-14T19%3A52%3A39IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_ieee_&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Automatic%20Report%20Generation%20for%20Chest%20X-Ray%20Images%20via%20Adversarial%20Reinforcement%20Learning&rft.jtitle=IEEE%20access&rft.au=Hou,%20Daibing&rft.date=2021-01-01&rft.volume=9&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.issn=2169-3536&rft.eissn=2169-3536&rft.coden=IAECCG&rft_id=info:doi/10.1109/ACCESS.2021.3056175&rft_dat=%3Cproquest_ieee_%3E2487437961%3C/proquest_ieee_%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2487437961&rft_id=info:pmid/&rft_ieee_id=9343868&rft_doaj_id=oai_doaj_org_article_bb739072aad0492b871a592e96848a55&rfr_iscdi=true