Towards Low-Resource Harmful Meme Detection with LMM Agents
The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an a...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Huang, Jianzhao Lin, Hongzhan Liu, Ziyan Luo, Ziyang Chen, Guang Ma, Jing |
description | The proliferation of Internet memes in the age of social media necessitates
effective identification of harmful ones. Due to the dynamic nature of memes,
existing data-driven models may struggle in low-resource scenarios where only a
few labeled examples are available. In this paper, we propose an agency-driven
framework for low-resource harmful meme detection, employing both outward and
inward analysis with few-shot annotated samples. Inspired by the powerful
capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first
retrieve relative memes with annotations to leverage label information as
auxiliary signals for the LMM agent. Then, we elicit knowledge-revising
behavior within the LMM agent to derive well-generalized insights into meme
harmfulness. By combining these strategies, our approach enables dialectical
reasoning over intricate and implicit harm-indicative patterns. Extensive
experiments conducted on three meme datasets demonstrate that our proposed
approach achieves superior performance than state-of-the-art methods on the
low-resource harmful meme detection task. |
doi_str_mv | 10.48550/arxiv.2411.05383 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2411_05383</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2411_05383</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2411_053833</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjE01DMwNbYw5mSwDskvTyxKKVbwyS_XDUotzi8tSk5V8Egsyk0rzVHwTc1NVXBJLUlNLsnMz1MozyzJUPDx9VVwTE_NKynmYWBNS8wpTuWF0twM8m6uIc4eumBr4guKMnMTiyrjQdbFg60zJqwCAH8FNCE</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><source>arXiv.org</source><creator>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</creator><creatorcontrib>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</creatorcontrib><description>The proliferation of Internet memes in the age of social media necessitates
effective identification of harmful ones. Due to the dynamic nature of memes,
existing data-driven models may struggle in low-resource scenarios where only a
few labeled examples are available. In this paper, we propose an agency-driven
framework for low-resource harmful meme detection, employing both outward and
inward analysis with few-shot annotated samples. Inspired by the powerful
capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first
retrieve relative memes with annotations to leverage label information as
auxiliary signals for the LMM agent. Then, we elicit knowledge-revising
behavior within the LMM agent to derive well-generalized insights into meme
harmfulness. By combining these strategies, our approach enables dialectical
reasoning over intricate and implicit harm-indicative patterns. Extensive
experiments conducted on three meme datasets demonstrate that our proposed
approach achieves superior performance than state-of-the-art methods on the
low-resource harmful meme detection task.</description><identifier>DOI: 10.48550/arxiv.2411.05383</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2024-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2411.05383$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2411.05383$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Huang, Jianzhao</creatorcontrib><creatorcontrib>Lin, Hongzhan</creatorcontrib><creatorcontrib>Liu, Ziyan</creatorcontrib><creatorcontrib>Luo, Ziyang</creatorcontrib><creatorcontrib>Chen, Guang</creatorcontrib><creatorcontrib>Ma, Jing</creatorcontrib><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><description>The proliferation of Internet memes in the age of social media necessitates
effective identification of harmful ones. Due to the dynamic nature of memes,
existing data-driven models may struggle in low-resource scenarios where only a
few labeled examples are available. In this paper, we propose an agency-driven
framework for low-resource harmful meme detection, employing both outward and
inward analysis with few-shot annotated samples. Inspired by the powerful
capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first
retrieve relative memes with annotations to leverage label information as
auxiliary signals for the LMM agent. Then, we elicit knowledge-revising
behavior within the LMM agent to derive well-generalized insights into meme
harmfulness. By combining these strategies, our approach enables dialectical
reasoning over intricate and implicit harm-indicative patterns. Extensive
experiments conducted on three meme datasets demonstrate that our proposed
approach achieves superior performance than state-of-the-art methods on the
low-resource harmful meme detection task.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjE01DMwNbYw5mSwDskvTyxKKVbwyS_XDUotzi8tSk5V8Egsyk0rzVHwTc1NVXBJLUlNLsnMz1MozyzJUPDx9VVwTE_NKynmYWBNS8wpTuWF0twM8m6uIc4eumBr4guKMnMTiyrjQdbFg60zJqwCAH8FNCE</recordid><startdate>20241108</startdate><enddate>20241108</enddate><creator>Huang, Jianzhao</creator><creator>Lin, Hongzhan</creator><creator>Liu, Ziyan</creator><creator>Luo, Ziyang</creator><creator>Chen, Guang</creator><creator>Ma, Jing</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241108</creationdate><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><author>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2411_053833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Huang, Jianzhao</creatorcontrib><creatorcontrib>Lin, Hongzhan</creatorcontrib><creatorcontrib>Liu, Ziyan</creatorcontrib><creatorcontrib>Luo, Ziyang</creatorcontrib><creatorcontrib>Chen, Guang</creatorcontrib><creatorcontrib>Ma, Jing</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Huang, Jianzhao</au><au>Lin, Hongzhan</au><au>Liu, Ziyan</au><au>Luo, Ziyang</au><au>Chen, Guang</au><au>Ma, Jing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Low-Resource Harmful Meme Detection with LMM Agents</atitle><date>2024-11-08</date><risdate>2024</risdate><abstract>The proliferation of Internet memes in the age of social media necessitates
effective identification of harmful ones. Due to the dynamic nature of memes,
existing data-driven models may struggle in low-resource scenarios where only a
few labeled examples are available. In this paper, we propose an agency-driven
framework for low-resource harmful meme detection, employing both outward and
inward analysis with few-shot annotated samples. Inspired by the powerful
capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first
retrieve relative memes with annotations to leverage label information as
auxiliary signals for the LMM agent. Then, we elicit knowledge-revising
behavior within the LMM agent to derive well-generalized insights into meme
harmfulness. By combining these strategies, our approach enables dialectical
reasoning over intricate and implicit harm-indicative patterns. Extensive
experiments conducted on three meme datasets demonstrate that our proposed
approach achieves superior performance than state-of-the-art methods on the
low-resource harmful meme detection task.</abstract><doi>10.48550/arxiv.2411.05383</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2411.05383 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2411_05383 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | Towards Low-Resource Harmful Meme Detection with LMM Agents |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T02%3A01%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Low-Resource%20Harmful%20Meme%20Detection%20with%20LMM%20Agents&rft.au=Huang,%20Jianzhao&rft.date=2024-11-08&rft_id=info:doi/10.48550/arxiv.2411.05383&rft_dat=%3Carxiv_GOX%3E2411_05383%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |