SemEval-2017 Task 3: Community Question Answering
SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year, we reran the four subtasks from SemEval-2016:(A) Question-Comment Similarity,(B) Question-Question Similarity,(C) Question-External Comment Similarity, and (D) Rerank the correct answers for a new question in Ar...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Nakov, Preslav Hoogeveen, Doris Màrquez, Lluís Moschitti, Alessandro Mubarak, Hamdy Baldwin, Timothy Verspoor, Karin |
description | SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year,
we reran the four subtasks from SemEval-2016:(A) Question-Comment
Similarity,(B) Question-Question Similarity,(C) Question-External Comment
Similarity, and (D) Rerank the correct answers for a new question in Arabic,
providing all the data from 2015 and 2016 for training, and fresh data for
testing. Additionally, we added a new subtask E in order to enable
experimentation with Multi-domain Question Duplicate Detection in a
larger-scale scenario, using StackExchange subforums. A total of 23 teams
participated in the task, and submitted a total of 85 runs (36 primary and 49
contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask
E. A variety of approaches and features were used by the participating systems
to address the different subtasks. The best systems achieved an official score
(MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D,
respectively. These scores are better than the baselines, especially for
subtasks A-C. |
doi_str_mv | 10.48550/arxiv.1912.00730 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1912_00730</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1912_00730</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-75cfef902a5980a03330bca904fd6e10019cb21a1edaa36ffe3f3bce5fd79c273</originalsourceid><addsrcrecordid>eNotzk1uwjAQQGFvWFTQA3RVXyDpjAfHmB2K6I-EVFXNPpo4Y2RBAkqAlttXpV293dOn1ANCPl9YC088fKdLjh5NDuAI7hR-Sre-8D4zgE5XPO40LXV56Lpzn05X_XGW8ZQOvV7145cMqd_O1CTyfpT7_05V9byuytds8_7yVq42GRcOMmdDlOjBsPULYCAiaAJ7mMe2EARAHxqDjNIyUxGjUKQmiI2t88E4mqrHv-3NXB-H1PFwrX_t9c1OP55LPY4</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>SemEval-2017 Task 3: Community Question Answering</title><source>arXiv.org</source><creator>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</creator><creatorcontrib>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</creatorcontrib><description>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year,
we reran the four subtasks from SemEval-2016:(A) Question-Comment
Similarity,(B) Question-Question Similarity,(C) Question-External Comment
Similarity, and (D) Rerank the correct answers for a new question in Arabic,
providing all the data from 2015 and 2016 for training, and fresh data for
testing. Additionally, we added a new subtask E in order to enable
experimentation with Multi-domain Question Duplicate Detection in a
larger-scale scenario, using StackExchange subforums. A total of 23 teams
participated in the task, and submitted a total of 85 runs (36 primary and 49
contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask
E. A variety of approaches and features were used by the participating systems
to address the different subtasks. The best systems achieved an official score
(MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D,
respectively. These scores are better than the baselines, especially for
subtasks A-C.</description><identifier>DOI: 10.48550/arxiv.1912.00730</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Information Retrieval ; Computer Science - Learning</subject><creationdate>2019-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1912.00730$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1912.00730$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Nakov, Preslav</creatorcontrib><creatorcontrib>Hoogeveen, Doris</creatorcontrib><creatorcontrib>Màrquez, Lluís</creatorcontrib><creatorcontrib>Moschitti, Alessandro</creatorcontrib><creatorcontrib>Mubarak, Hamdy</creatorcontrib><creatorcontrib>Baldwin, Timothy</creatorcontrib><creatorcontrib>Verspoor, Karin</creatorcontrib><title>SemEval-2017 Task 3: Community Question Answering</title><description>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year,
we reran the four subtasks from SemEval-2016:(A) Question-Comment
Similarity,(B) Question-Question Similarity,(C) Question-External Comment
Similarity, and (D) Rerank the correct answers for a new question in Arabic,
providing all the data from 2015 and 2016 for training, and fresh data for
testing. Additionally, we added a new subtask E in order to enable
experimentation with Multi-domain Question Duplicate Detection in a
larger-scale scenario, using StackExchange subforums. A total of 23 teams
participated in the task, and submitted a total of 85 runs (36 primary and 49
contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask
E. A variety of approaches and features were used by the participating systems
to address the different subtasks. The best systems achieved an official score
(MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D,
respectively. These scores are better than the baselines, especially for
subtasks A-C.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Information Retrieval</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzk1uwjAQQGFvWFTQA3RVXyDpjAfHmB2K6I-EVFXNPpo4Y2RBAkqAlttXpV293dOn1ANCPl9YC088fKdLjh5NDuAI7hR-Sre-8D4zgE5XPO40LXV56Lpzn05X_XGW8ZQOvV7145cMqd_O1CTyfpT7_05V9byuytds8_7yVq42GRcOMmdDlOjBsPULYCAiaAJ7mMe2EARAHxqDjNIyUxGjUKQmiI2t88E4mqrHv-3NXB-H1PFwrX_t9c1OP55LPY4</recordid><startdate>20191202</startdate><enddate>20191202</enddate><creator>Nakov, Preslav</creator><creator>Hoogeveen, Doris</creator><creator>Màrquez, Lluís</creator><creator>Moschitti, Alessandro</creator><creator>Mubarak, Hamdy</creator><creator>Baldwin, Timothy</creator><creator>Verspoor, Karin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20191202</creationdate><title>SemEval-2017 Task 3: Community Question Answering</title><author>Nakov, Preslav ; Hoogeveen, Doris ; Màrquez, Lluís ; Moschitti, Alessandro ; Mubarak, Hamdy ; Baldwin, Timothy ; Verspoor, Karin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-75cfef902a5980a03330bca904fd6e10019cb21a1edaa36ffe3f3bce5fd79c273</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Information Retrieval</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Nakov, Preslav</creatorcontrib><creatorcontrib>Hoogeveen, Doris</creatorcontrib><creatorcontrib>Màrquez, Lluís</creatorcontrib><creatorcontrib>Moschitti, Alessandro</creatorcontrib><creatorcontrib>Mubarak, Hamdy</creatorcontrib><creatorcontrib>Baldwin, Timothy</creatorcontrib><creatorcontrib>Verspoor, Karin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Nakov, Preslav</au><au>Hoogeveen, Doris</au><au>Màrquez, Lluís</au><au>Moschitti, Alessandro</au><au>Mubarak, Hamdy</au><au>Baldwin, Timothy</au><au>Verspoor, Karin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SemEval-2017 Task 3: Community Question Answering</atitle><date>2019-12-02</date><risdate>2019</risdate><abstract>SemEval-2017 We describe SemEval-2017 Task 3 on Community Question Answering. This year,
we reran the four subtasks from SemEval-2016:(A) Question-Comment
Similarity,(B) Question-Question Similarity,(C) Question-External Comment
Similarity, and (D) Rerank the correct answers for a new question in Arabic,
providing all the data from 2015 and 2016 for training, and fresh data for
testing. Additionally, we added a new subtask E in order to enable
experimentation with Multi-domain Question Duplicate Detection in a
larger-scale scenario, using StackExchange subforums. A total of 23 teams
participated in the task, and submitted a total of 85 runs (36 primary and 49
contrastive) for subtasks A-D. Unfortunately, no teams participated in subtask
E. A variety of approaches and features were used by the participating systems
to address the different subtasks. The best systems achieved an official score
(MAP) of 88.43, 47.22, 15.46, and 61.16 in subtasks A, B, C, and D,
respectively. These scores are better than the baselines, especially for
subtasks A-C.</abstract><doi>10.48550/arxiv.1912.00730</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1912.00730 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1912_00730 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computation and Language Computer Science - Information Retrieval Computer Science - Learning |
title | SemEval-2017 Task 3: Community Question Answering |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-24T18%3A56%3A00IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SemEval-2017%20Task%203:%20Community%20Question%20Answering&rft.au=Nakov,%20Preslav&rft.date=2019-12-02&rft_id=info:doi/10.48550/arxiv.1912.00730&rft_dat=%3Carxiv_GOX%3E1912_00730%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |