SemEval-2017 Task 3: Community Question Answering

SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Ar...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Nakov, Preslav, Hoogeveen, Doris, Màrquez, Lluís, Moschitti, Alessandro, Mubarak, Hamdy, Baldwin, Timothy, Verspoor, Karin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Nakov, Preslav
Hoogeveen, Doris
Màrquez, Lluís
Moschitti, Alessandro
Mubarak, Hamdy
Baldwin, Timothy
Verspoor, Karin
description SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Arabic, providing all the data from 2015 and 2016 for training, and fresh data for testing. Additionally, we added a new subtask E in order to enable experimentation with Multi-domain Question Duplicate Detection in a larger-scale scenario, using StackExchange subforums. A total of 23 teams participated in the task, and submitted a total of 85 runs (36 primary and 49 contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask E. A variety of approaches and features were used by the participating systems to address the different subtasks. The best systems achieved an official score (MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D, respectively. These scores are better than the baselines, especially for subtasks A-C.
doi_str_mv 10.48550/arxiv.1912.00730
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1912_00730</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1912_00730</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-75cfef902a5980a03330bca904fd6e10019cb21a1edaa36ffe3f3bce5fd79c273</originalsourceid><addsrcrecordid>eNotzk1uwjAQQGFvWFTQA3RVXyDpjAfHmB2K6I-EVFXNPpo4Y2RBAkqAlttXpV293dOn1ANCPl9YC088fKdLjh5NDuAI7hR-Sre-8D4zgE5XPO40LXV56Lpzn05X_XGW8ZQOvV7145cMqd_O1CTyfpT7_05V9byuytds8_7yVq42GRcOMmdDlOjBsPULYCAiaAJ7mMe2EARAHxqDjNIyUxGjUKQmiI2t88E4mqrHv-3NXB-H1PFwrX_t9c1OP55LPY4</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>SemEval-2017 Task 3: Community Question Answering</title><source>arXiv.org</source><creator>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</creator><creatorcontrib>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</creatorcontrib><description>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Arabic, providing all the data from 2015 and 2016 for training, and fresh data for testing. Additionally, we added a new subtask E in order to enable experimentation with Multi-domain Question Duplicate Detection in a larger-scale scenario, using StackExchange subforums. A total of 23 teams participated in the task, and submitted a total of 85 runs (36 primary and 49 contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask E. A variety of approaches and features were used by the participating systems to address the different subtasks. The best systems achieved an official score (MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D, respectively. These scores are better than the baselines, especially for subtasks A-C.</description><identifier>DOI: 10.48550/arxiv.1912.00730</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Information Retrieval ; Computer Science - Learning</subject><creationdate>2019-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1912.00730$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1912.00730$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Nakov, Preslav</creatorcontrib><creatorcontrib>Hoogeveen, Doris</creatorcontrib><creatorcontrib>Màrquez, Lluís</creatorcontrib><creatorcontrib>Moschitti, Alessandro</creatorcontrib><creatorcontrib>Mubarak, Hamdy</creatorcontrib><creatorcontrib>Baldwin, Timothy</creatorcontrib><creatorcontrib>Verspoor, Karin</creatorcontrib><title>SemEval-2017 Task 3: Community Question Answering</title><description>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Arabic, providing all the data from 2015 and 2016 for training, and fresh data for testing. Additionally, we added a new subtask E in order to enable experimentation with Multi-domain Question Duplicate Detection in a larger-scale scenario, using StackExchange subforums. A total of 23 teams participated in the task, and submitted a total of 85 runs (36 primary and 49 contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask E. A variety of approaches and features were used by the participating systems to address the different subtasks. The best systems achieved an official score (MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D, respectively. These scores are better than the baselines, especially for subtasks A-C.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Information Retrieval</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzk1uwjAQQGFvWFTQA3RVXyDpjAfHmB2K6I-EVFXNPpo4Y2RBAkqAlttXpV293dOn1ANCPl9YC088fKdLjh5NDuAI7hR-Sre-8D4zgE5XPO40LXV56Lpzn05X_XGW8ZQOvV7145cMqd_O1CTyfpT7_05V9byuytds8_7yVq42GRcOMmdDlOjBsPULYCAiaAJ7mMe2EARAHxqDjNIyUxGjUKQmiI2t88E4mqrHv-3NXB-H1PFwrX_t9c1OP55LPY4</recordid><startdate>20191202</startdate><enddate>20191202</enddate><creator>Nakov, Preslav</creator><creator>Hoogeveen, Doris</creator><creator>Màrquez, Lluís</creator><creator>Moschitti, Alessandro</creator><creator>Mubarak, Hamdy</creator><creator>Baldwin, Timothy</creator><creator>Verspoor, Karin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20191202</creationdate><title>SemEval-2017 Task 3: Community Question Answering</title><author>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-75cfef902a5980a03330bca904fd6e10019cb21a1edaa36ffe3f3bce5fd79c273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Information Retrieval</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Nakov, Preslav</creatorcontrib><creatorcontrib>Hoogeveen, Doris</creatorcontrib><creatorcontrib>Màrquez, Lluís</creatorcontrib><creatorcontrib>Moschitti, Alessandro</creatorcontrib><creatorcontrib>Mubarak, Hamdy</creatorcontrib><creatorcontrib>Baldwin, Timothy</creatorcontrib><creatorcontrib>Verspoor, Karin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Nakov, Preslav</au><au>Hoogeveen, Doris</au><au>Màrquez, Lluís</au><au>Moschitti, Alessandro</au><au>Mubarak, Hamdy</au><au>Baldwin, Timothy</au><au>Verspoor, Karin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SemEval-2017 Task 3: Community Question Answering</atitle><date>2019-12-02</date><risdate>2019</risdate><abstract>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Arabic, providing all the data from 2015 and 2016 for training, and fresh data for testing. Additionally, we added a new subtask E in order to enable experimentation with Multi-domain Question Duplicate Detection in a larger-scale scenario, using StackExchange subforums. A total of 23 teams participated in the task, and submitted a total of 85 runs (36 primary and 49 contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask E. A variety of approaches and features were used by the participating systems to address the different subtasks. The best systems achieved an official score (MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D, respectively. These scores are better than the baselines, especially for subtasks A-C.</abstract><doi>10.48550/arxiv.1912.00730</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1912.00730
ispartof
issn
language eng
recordid cdi_arxiv_primary_1912_00730
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
Computer Science - Information Retrieval
Computer Science - Learning
title SemEval-2017 Task 3: Community Question Answering
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-24T18%3A56%3A00IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SemEval-2017%20Task%203:%20Community%20Question%20Answering&rft.au=Nakov,%20Preslav&rft.date=2019-12-02&rft_id=info:doi/10.48550/arxiv.1912.00730&rft_dat=%3Carxiv_GOX%3E1912_00730%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true