Entity-aware answer sentence selection for question answering with transformer-based language models
The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of...
Gespeichert in:
Veröffentlicht in: | Journal of intelligent information systems 2022-12, Vol.59 (3), p.755-777 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 777 |
---|---|
container_issue | 3 |
container_start_page | 755 |
container_title | Journal of intelligent information systems |
container_volume | 59 |
creator | Abbasiantaeb, Zahra Momtazi, Saeedeh |
description | The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively. |
doi_str_mv | 10.1007/s10844-022-00724-6 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2737266372</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2737266372</sourcerecordid><originalsourceid>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</originalsourceid><addsrcrecordid>eNp9UE1LAzEQDaJgrf4BTwHP0WQ2u0mOUuoHFLzoOWSzSd2yzdYkpfTfG7uCNy_zxXtvZh5Ct4zeM0rFQ2JUck4oACktcNKcoRmrRUVEI-pzNKMKaqIUhUt0ldKGUqpkQ2eoW4bc5yMxBxMdNiEdXMTJheyCdaUYnM39GLAfI_7au3RqJlgf1vjQ50-cYxkUwNZF0prkOjyYsN6btcPbsXNDukYX3gzJ3fzmOfp4Wr4vXsjq7fl18bgiFiRkwlppGPeGCejA2LayrZXGci4Mtx46zyX30PJKMsm44rJkwetOtU3LPKhqju4m3V0cT8fqzbiPoazUICoBTVNCQcGEsnFMKTqvd7HfmnjUjOofN_Xkpi5u6pObuimkaiKl3c_jLv5J_8P6BoE-eZw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2737266372</pqid></control><display><type>article</type><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><source>SpringerLink Journals - AutoHoldings</source><creator>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</creator><creatorcontrib>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</creatorcontrib><description>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</description><identifier>ISSN: 0925-9902</identifier><identifier>EISSN: 1573-7675</identifier><identifier>DOI: 10.1007/s10844-022-00724-6</identifier><language>eng</language><publisher>New York: Springer US</publisher><subject>Artificial Intelligence ; Artificial neural networks ; Computer Science ; Data Structures and Information Theory ; Datasets ; Information Storage and Retrieval ; Information systems ; IT in Business ; Language ; Matching ; Natural Language Processing (NLP) ; Neural networks ; Probability ; Questions ; Reading comprehension ; Transformers</subject><ispartof>Journal of intelligent information systems, 2022-12, Vol.59 (3), p.755-777</ispartof><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022</rights><rights>The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2022.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</citedby><cites>FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</cites><orcidid>0000-0002-8110-1342</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10844-022-00724-6$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10844-022-00724-6$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,776,780,27901,27902,41464,42533,51294</link.rule.ids></links><search><creatorcontrib>Abbasiantaeb, Zahra</creatorcontrib><creatorcontrib>Momtazi, Saeedeh</creatorcontrib><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><title>Journal of intelligent information systems</title><addtitle>J Intell Inf Syst</addtitle><description>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</description><subject>Artificial Intelligence</subject><subject>Artificial neural networks</subject><subject>Computer Science</subject><subject>Data Structures and Information Theory</subject><subject>Datasets</subject><subject>Information Storage and Retrieval</subject><subject>Information systems</subject><subject>IT in Business</subject><subject>Language</subject><subject>Matching</subject><subject>Natural Language Processing (NLP)</subject><subject>Neural networks</subject><subject>Probability</subject><subject>Questions</subject><subject>Reading comprehension</subject><subject>Transformers</subject><issn>0925-9902</issn><issn>1573-7675</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNp9UE1LAzEQDaJgrf4BTwHP0WQ2u0mOUuoHFLzoOWSzSd2yzdYkpfTfG7uCNy_zxXtvZh5Ct4zeM0rFQ2JUck4oACktcNKcoRmrRUVEI-pzNKMKaqIUhUt0ldKGUqpkQ2eoW4bc5yMxBxMdNiEdXMTJheyCdaUYnM39GLAfI_7au3RqJlgf1vjQ50-cYxkUwNZF0prkOjyYsN6btcPbsXNDukYX3gzJ3fzmOfp4Wr4vXsjq7fl18bgiFiRkwlppGPeGCejA2LayrZXGci4Mtx46zyX30PJKMsm44rJkwetOtU3LPKhqju4m3V0cT8fqzbiPoazUICoBTVNCQcGEsnFMKTqvd7HfmnjUjOofN_Xkpi5u6pObuimkaiKl3c_jLv5J_8P6BoE-eZw</recordid><startdate>20221201</startdate><enddate>20221201</enddate><creator>Abbasiantaeb, Zahra</creator><creator>Momtazi, Saeedeh</creator><general>Springer US</general><general>Springer Nature B.V</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SC</scope><scope>7WY</scope><scope>7WZ</scope><scope>7XB</scope><scope>87Z</scope><scope>8AL</scope><scope>8FD</scope><scope>8FE</scope><scope>8FG</scope><scope>8FK</scope><scope>8FL</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>ARAPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BEZIV</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>FRNLG</scope><scope>F~G</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>JQ2</scope><scope>K60</scope><scope>K6~</scope><scope>K7-</scope><scope>L.-</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><scope>M0C</scope><scope>M0N</scope><scope>P5Z</scope><scope>P62</scope><scope>PQBIZ</scope><scope>PQBZA</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>Q9U</scope><orcidid>https://orcid.org/0000-0002-8110-1342</orcidid></search><sort><creationdate>20221201</creationdate><title>Entity-aware answer sentence selection for question answering with transformer-based language models</title><author>Abbasiantaeb, Zahra ; Momtazi, Saeedeh</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c282t-1b8a14fa172d2acb3cbc8ac447a4cf2df484f2b4381814948818745d9b6b1f293</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Artificial Intelligence</topic><topic>Artificial neural networks</topic><topic>Computer Science</topic><topic>Data Structures and Information Theory</topic><topic>Datasets</topic><topic>Information Storage and Retrieval</topic><topic>Information systems</topic><topic>IT in Business</topic><topic>Language</topic><topic>Matching</topic><topic>Natural Language Processing (NLP)</topic><topic>Neural networks</topic><topic>Probability</topic><topic>Questions</topic><topic>Reading comprehension</topic><topic>Transformers</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Abbasiantaeb, Zahra</creatorcontrib><creatorcontrib>Momtazi, Saeedeh</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Computer and Information Systems Abstracts</collection><collection>ABI/INFORM Complete</collection><collection>ABI/INFORM Global (PDF only)</collection><collection>ProQuest Central (purchase pre-March 2016)</collection><collection>ABI/INFORM Global (Alumni Edition)</collection><collection>Computing Database (Alumni Edition)</collection><collection>Technology Research Database</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ABI/INFORM Collection (Alumni Edition)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>Advanced Technologies & Aerospace Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Business Premium Collection</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>Business Premium Collection (Alumni)</collection><collection>ABI/INFORM Global (Corporate)</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Computer Science Collection</collection><collection>ProQuest Business Collection (Alumni Edition)</collection><collection>ProQuest Business Collection</collection><collection>Computer Science Database</collection><collection>ABI/INFORM Professional Advanced</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><collection>ABI/INFORM Global</collection><collection>Computing Database</collection><collection>Advanced Technologies & Aerospace Database</collection><collection>ProQuest Advanced Technologies & Aerospace Collection</collection><collection>One Business (ProQuest)</collection><collection>ProQuest One Business (Alumni)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central Basic</collection><jtitle>Journal of intelligent information systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Abbasiantaeb, Zahra</au><au>Momtazi, Saeedeh</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Entity-aware answer sentence selection for question answering with transformer-based language models</atitle><jtitle>Journal of intelligent information systems</jtitle><stitle>J Intell Inf Syst</stitle><date>2022-12-01</date><risdate>2022</risdate><volume>59</volume><issue>3</issue><spage>755</spage><epage>777</epage><pages>755-777</pages><issn>0925-9902</issn><eissn>1573-7675</eissn><abstract>The Answer Sentence Selection (AS2) task is defined as the task of ranking the candidate answers for each question based on a matching score. The matching score is the probability of being a correct answer for a given question. Detecting the question class and matching it with the named entities of the answer sentence to narrow down the search space was used in primary question answering systems. We used this idea in the state-of-the-art text matching models namely, Transformer-based language models. In this paper, we proposed two different architectures: Ent-match and Ent-add, while using two different question classifiers: Convolutional Neural Network-based (CNN-based) and rule-based. The proposed models outperform the state-of-the-art AS2 model, namely TANDA and RoBERTa-base on both TREC-QA and Wiki-QA datasets. Using Wiki-QA, the Ent-add (CNN-based) model outperforms the TANDA model by 2.1% and 1.9% improvement over Mean Average Precision (MAP) and Mean Reciprocal Rank (MRR) metrics, respectively. Over the TREC-QA dataset the Ent-match (CNN-based) model outperformed the TANDA model with 1.5% and 1.4% improvement over MAP and MRR, respectively.</abstract><cop>New York</cop><pub>Springer US</pub><doi>10.1007/s10844-022-00724-6</doi><tpages>23</tpages><orcidid>https://orcid.org/0000-0002-8110-1342</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0925-9902 |
ispartof | Journal of intelligent information systems, 2022-12, Vol.59 (3), p.755-777 |
issn | 0925-9902 1573-7675 |
language | eng |
recordid | cdi_proquest_journals_2737266372 |
source | SpringerLink Journals - AutoHoldings |
subjects | Artificial Intelligence Artificial neural networks Computer Science Data Structures and Information Theory Datasets Information Storage and Retrieval Information systems IT in Business Language Matching Natural Language Processing (NLP) Neural networks Probability Questions Reading comprehension Transformers |
title | Entity-aware answer sentence selection for question answering with transformer-based language models |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-29T17%3A51%3A21IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Entity-aware%20answer%20sentence%20selection%20for%20question%20answering%20with%20transformer-based%20language%20models&rft.jtitle=Journal%20of%20intelligent%20information%20systems&rft.au=Abbasiantaeb,%20Zahra&rft.date=2022-12-01&rft.volume=59&rft.issue=3&rft.spage=755&rft.epage=777&rft.pages=755-777&rft.issn=0925-9902&rft.eissn=1573-7675&rft_id=info:doi/10.1007/s10844-022-00724-6&rft_dat=%3Cproquest_cross%3E2737266372%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2737266372&rft_id=info:pmid/&rfr_iscdi=true |