Metaphor Understanding Challenge Dataset for LLMs
Metaphors in natural language are a reflection of fundamental cognitive processes such as analogical reasoning and categorisation, and are deeply rooted in everyday communication. Metaphor understanding is therefore an essential task for large language models (LLMs). We release the Metaphor Understa...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Tong, Xiaoyu Choenni, Rochelle Lewis, Martha Shutova, Ekaterina |
description | Metaphors in natural language are a reflection of fundamental cognitive
processes such as analogical reasoning and categorisation, and are deeply
rooted in everyday communication. Metaphor understanding is therefore an
essential task for large language models (LLMs). We release the Metaphor
Understanding Challenge Dataset (MUNCH), designed to evaluate the metaphor
understanding capabilities of LLMs. The dataset provides over 10k paraphrases
for sentences containing metaphor use, as well as 1.5k instances containing
inapt paraphrases. The inapt paraphrases were carefully selected to serve as
control to determine whether the model indeed performs full metaphor
interpretation or rather resorts to lexical similarity. All apt and inapt
paraphrases were manually annotated. The metaphorical sentences cover natural
metaphor uses across 4 genres (academic, news, fiction, and conversation), and
they exhibit different levels of novelty. Experiments with LLaMA and GPT-3.5
demonstrate that MUNCH presents a challenging task for LLMs. The dataset is
freely accessible at
https://github.com/xiaoyuisrain/metaphor-understanding-challenge. |
doi_str_mv | 10.48550/arxiv.2403.11810 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2403_11810</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2403_11810</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-c9fcf037696d42531034f375bc27cfd2bb8b7e2fdde5b42061b7436c694495b33</originalsourceid><addsrcrecordid>eNotzr1OwzAYhWEvHVDpBTCRG0hq-_NPMqK0BaRELGWOPv-1kYIbOVHV3j0QmM7y6ugh5InRQpRS0i2mW38tuKBQMFYy-kBY62ccz5eUfUbn0zRjdH08ZfUZh8HHk892OOPk5yz8NE3TTo9kFXCY_OZ_1-R42B_rt7z5eH2vX5oclaa5rYINFLSqlBNcAqMgAmhpLNc2OG5MabTnwTkvjeBUMaMFKKsqISppANbk-e92MXdj6r8w3btfe7fY4RtA_j2U</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Metaphor Understanding Challenge Dataset for LLMs</title><source>arXiv.org</source><creator>Tong, Xiaoyu ; Choenni, Rochelle ; Lewis, Martha ; Shutova, Ekaterina</creator><creatorcontrib>Tong, Xiaoyu ; Choenni, Rochelle ; Lewis, Martha ; Shutova, Ekaterina</creatorcontrib><description>Metaphors in natural language are a reflection of fundamental cognitive
processes such as analogical reasoning and categorisation, and are deeply
rooted in everyday communication. Metaphor understanding is therefore an
essential task for large language models (LLMs). We release the Metaphor
Understanding Challenge Dataset (MUNCH), designed to evaluate the metaphor
understanding capabilities of LLMs. The dataset provides over 10k paraphrases
for sentences containing metaphor use, as well as 1.5k instances containing
inapt paraphrases. The inapt paraphrases were carefully selected to serve as
control to determine whether the model indeed performs full metaphor
interpretation or rather resorts to lexical similarity. All apt and inapt
paraphrases were manually annotated. The metaphorical sentences cover natural
metaphor uses across 4 genres (academic, news, fiction, and conversation), and
they exhibit different levels of novelty. Experiments with LLaMA and GPT-3.5
demonstrate that MUNCH presents a challenging task for LLMs. The dataset is
freely accessible at
https://github.com/xiaoyuisrain/metaphor-understanding-challenge.</description><identifier>DOI: 10.48550/arxiv.2403.11810</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2024-03</creationdate><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2403.11810$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2403.11810$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Tong, Xiaoyu</creatorcontrib><creatorcontrib>Choenni, Rochelle</creatorcontrib><creatorcontrib>Lewis, Martha</creatorcontrib><creatorcontrib>Shutova, Ekaterina</creatorcontrib><title>Metaphor Understanding Challenge Dataset for LLMs</title><description>Metaphors in natural language are a reflection of fundamental cognitive
processes such as analogical reasoning and categorisation, and are deeply
rooted in everyday communication. Metaphor understanding is therefore an
essential task for large language models (LLMs). We release the Metaphor
Understanding Challenge Dataset (MUNCH), designed to evaluate the metaphor
understanding capabilities of LLMs. The dataset provides over 10k paraphrases
for sentences containing metaphor use, as well as 1.5k instances containing
inapt paraphrases. The inapt paraphrases were carefully selected to serve as
control to determine whether the model indeed performs full metaphor
interpretation or rather resorts to lexical similarity. All apt and inapt
paraphrases were manually annotated. The metaphorical sentences cover natural
metaphor uses across 4 genres (academic, news, fiction, and conversation), and
they exhibit different levels of novelty. Experiments with LLaMA and GPT-3.5
demonstrate that MUNCH presents a challenging task for LLMs. The dataset is
freely accessible at
https://github.com/xiaoyuisrain/metaphor-understanding-challenge.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzr1OwzAYhWEvHVDpBTCRG0hq-_NPMqK0BaRELGWOPv-1kYIbOVHV3j0QmM7y6ugh5InRQpRS0i2mW38tuKBQMFYy-kBY62ccz5eUfUbn0zRjdH08ZfUZh8HHk892OOPk5yz8NE3TTo9kFXCY_OZ_1-R42B_rt7z5eH2vX5oclaa5rYINFLSqlBNcAqMgAmhpLNc2OG5MabTnwTkvjeBUMaMFKKsqISppANbk-e92MXdj6r8w3btfe7fY4RtA_j2U</recordid><startdate>20240318</startdate><enddate>20240318</enddate><creator>Tong, Xiaoyu</creator><creator>Choenni, Rochelle</creator><creator>Lewis, Martha</creator><creator>Shutova, Ekaterina</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240318</creationdate><title>Metaphor Understanding Challenge Dataset for LLMs</title><author>Tong, Xiaoyu ; Choenni, Rochelle ; Lewis, Martha ; Shutova, Ekaterina</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-c9fcf037696d42531034f375bc27cfd2bb8b7e2fdde5b42061b7436c694495b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Tong, Xiaoyu</creatorcontrib><creatorcontrib>Choenni, Rochelle</creatorcontrib><creatorcontrib>Lewis, Martha</creatorcontrib><creatorcontrib>Shutova, Ekaterina</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Tong, Xiaoyu</au><au>Choenni, Rochelle</au><au>Lewis, Martha</au><au>Shutova, Ekaterina</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Metaphor Understanding Challenge Dataset for LLMs</atitle><date>2024-03-18</date><risdate>2024</risdate><abstract>Metaphors in natural language are a reflection of fundamental cognitive
processes such as analogical reasoning and categorisation, and are deeply
rooted in everyday communication. Metaphor understanding is therefore an
essential task for large language models (LLMs). We release the Metaphor
Understanding Challenge Dataset (MUNCH), designed to evaluate the metaphor
understanding capabilities of LLMs. The dataset provides over 10k paraphrases
for sentences containing metaphor use, as well as 1.5k instances containing
inapt paraphrases. The inapt paraphrases were carefully selected to serve as
control to determine whether the model indeed performs full metaphor
interpretation or rather resorts to lexical similarity. All apt and inapt
paraphrases were manually annotated. The metaphorical sentences cover natural
metaphor uses across 4 genres (academic, news, fiction, and conversation), and
they exhibit different levels of novelty. Experiments with LLaMA and GPT-3.5
demonstrate that MUNCH presents a challenging task for LLMs. The dataset is
freely accessible at
https://github.com/xiaoyuisrain/metaphor-understanding-challenge.</abstract><doi>10.48550/arxiv.2403.11810</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2403.11810 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2403_11810 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | Metaphor Understanding Challenge Dataset for LLMs |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-21T09%3A05%3A42IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Metaphor%20Understanding%20Challenge%20Dataset%20for%20LLMs&rft.au=Tong,%20Xiaoyu&rft.date=2024-03-18&rft_id=info:doi/10.48550/arxiv.2403.11810&rft_dat=%3Carxiv_GOX%3E2403_11810%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |