Entity-aware answer sentence selection for question answering with transformer-based language models

The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of intelligent information systems 2022-12, Vol.59 (3), p.755-777
Hauptverfasser: Abbasiantaeb, Zahra, Momtazi, Saeedeh
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 777
container_issue 3
container_start_page 755
container_title Journal of intelligent information systems
container_volume 59
creator Abbasiantaeb, Zahra
Momtazi, Saeedeh
description The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.
doi_str_mv 10.1007/s10844-022-00724-6
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2737266372</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2737266372</sourcerecordid><originalsourceid>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</originalsourceid><addsrcrecordid>eNp9UE1LAzEQDaJgrf4BTwHP0WQ2u0mOUuoHFLzoOWSzSd2yzdYkpfTfG7uCNy_zxXtvZh5Ct4zeM0rFQ2JUck4oACktcNKcoRmrRUVEI-pzNKMKaqIUhUt0ldKGUqpkQ2eoW4bc5yMxBxMdNiEdXMTJheyCdaUYnM39GLAfI_7au3RqJlgf1vjQ50-cYxkUwNZF0prkOjyYsN6btcPbsXNDukYX3gzJ3fzmOfp4Wr4vXsjq7fl18bgiFiRkwlppGPeGCejA2LayrZXGci4Mtx46zyX30PJKMsm44rJkwetOtU3LPKhqju4m3V0cT8fqzbiPoazUICoBTVNCQcGEsnFMKTqvd7HfmnjUjOofN_Xkpi5u6pObuimkaiKl3c_jLv5J_8P6BoE-eZw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2737266372</pqid></control><display><type>article</type><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><source>SpringerLink Journals - AutoHoldings</source><creator>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</creator><creatorcontrib>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</creatorcontrib><description>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</description><identifier>ISSN: 0925-9902</identifier><identifier>EISSN: 1573-7675</identifier><identifier>DOI: 10.1007/s10844-022-00724-6</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial Intelligence ; Artificial neural networks ; Computer Science ; Data Structures and Information Theory ; Datasets ; Information Storage and Retrieval ; Information systems ; IT in Business ; Language ; Matching ; Natural Language Processing (NLP) ; Neural networks ; Probability ; Questions ; Reading comprehension ; Transformers</subject><ispartof>Journal of intelligent information systems, 2022-12, Vol.59 (3), p.755-777</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</citedby><cites>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</cites><orcidid>0000-0002-8110-1342</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10844-022-00724-6$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10844-022-00724-6$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Abbasiantaeb, Zahra</creatorcontrib><creatorcontrib>Momtazi, Saeedeh</creatorcontrib><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><title>Journal of intelligent information systems</title><addtitle>J Intell Inf Syst</addtitle><description>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</description><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Datasets</subject><subject>Information Storage and Retrieval</subject><subject>Information systems</subject><subject>IT in Business</subject><subject>Language</subject><subject>Matching</subject><subject>Natural Language Processing (NLP)</subject><subject>Neural networks</subject><subject>Probability</subject><subject>Questions</subject><subject>Reading comprehension</subject><subject>Transformers</subject><issn>0925-9902</issn><issn>1573-7675</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9UE1LAzEQDaJgrf4BTwHP0WQ2u0mOUuoHFLzoOWSzSd2yzdYkpfTfG7uCNy_zxXtvZh5Ct4zeM0rFQ2JUck4oACktcNKcoRmrRUVEI-pzNKMKaqIUhUt0ldKGUqpkQ2eoW4bc5yMxBxMdNiEdXMTJheyCdaUYnM39GLAfI_7au3RqJlgf1vjQ50-cYxkUwNZF0prkOjyYsN6btcPbsXNDukYX3gzJ3fzmOfp4Wr4vXsjq7fl18bgiFiRkwlppGPeGCejA2LayrZXGci4Mtx46zyX30PJKMsm44rJkwetOtU3LPKhqju4m3V0cT8fqzbiPoazUICoBTVNCQcGEsnFMKTqvd7HfmnjUjOofN_Xkpi5u6pObuimkaiKl3c_jLv5J_8P6BoE-eZw</recordid><startdate>20221201</startdate><enddate>20221201</enddate><creator>Abbasiantaeb, Zahra</creator><creator>Momtazi, Saeedeh</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-8110-1342</orcidid></search><sort><creationdate>20221201</creationdate><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><author>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Datasets</topic><topic>Information Storage and Retrieval</topic><topic>Information systems</topic><topic>IT in Business</topic><topic>Language</topic><topic>Matching</topic><topic>Natural Language Processing (NLP)</topic><topic>Neural networks</topic><topic>Probability</topic><topic>Questions</topic><topic>Reading comprehension</topic><topic>Transformers</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Abbasiantaeb, Zahra</creatorcontrib><creatorcontrib>Momtazi, Saeedeh</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Complete</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies &amp; Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Advanced Technologies &amp; Aerospace Database</collection><collection>ProQuest Advanced Technologies &amp; Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Journal of intelligent information systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Abbasiantaeb, Zahra</au><au>Momtazi, Saeedeh</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Entity-aware answer sentence selection for question answering with transformer-based language models</atitle><jtitle>Journal of intelligent information systems</jtitle><stitle>J Intell Inf Syst</stitle><date>2022-12-01</date><risdate>2022</risdate><volume>59</volume><issue>3</issue><spage>755</spage><epage>777</epage><pages>755-777</pages><issn>0925-9902</issn><eissn>1573-7675</eissn><abstract>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10844-022-00724-6</doi><tpages>23</tpages><orcidid>https://orcid.org/0000-0002-8110-1342</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0925-9902
ispartof Journal of intelligent information systems, 2022-12, Vol.59 (3), p.755-777
issn 0925-9902
1573-7675
language eng
recordid cdi_proquest_journals_2737266372
source SpringerLink Journals - AutoHoldings
subjects Artificial Intelligence
Artificial neural networks
Computer Science
Data Structures and Information Theory
Datasets
Information Storage and Retrieval
Information systems
IT in Business
Language
Matching
Natural Language Processing (NLP)
Neural networks
Probability
Questions
Reading comprehension
Transformers
title Entity-aware answer sentence selection for question answering with transformer-based language models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T17%3A51%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Entity-aware%20answer%20sentence%20selection%20for%20question%20answering%20with%20transformer-based%20language%20models&rft.jtitle=Journal%20of%20intelligent%20information%20systems&rft.au=Abbasiantaeb,%20Zahra&rft.date=2022-12-01&rft.volume=59&rft.issue=3&rft.spage=755&rft.epage=777&rft.pages=755-777&rft.issn=0925-9902&rft.eissn=1573-7675&rft_id=info:doi/10.1007/s10844-022-00724-6&rft_dat=%3Cproquest_cross%3E2737266372%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2737266372&rft_id=info:pmid/&rfr_iscdi=true