Towards Low-Resource Harmful Meme Detection with LMM Agents

The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Huang, Jianzhao, Lin, Hongzhan, Liu, Ziyan, Luo, Ziyang, Chen, Guang, Ma, Jing
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Huang, Jianzhao
Lin, Hongzhan
Liu, Ziyan
Luo, Ziyang
Chen, Guang
Ma, Jing
description The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an agency-driven framework for low-resource harmful meme detection, employing both outward and inward analysis with few-shot annotated samples. Inspired by the powerful capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first retrieve relative memes with annotations to leverage label information as auxiliary signals for the LMM agent. Then, we elicit knowledge-revising behavior within the LMM agent to derive well-generalized insights into meme harmfulness. By combining these strategies, our approach enables dialectical reasoning over intricate and implicit harm-indicative patterns. Extensive experiments conducted on three meme datasets demonstrate that our proposed approach achieves superior performance than state-of-the-art methods on the low-resource harmful meme detection task.
doi_str_mv 10.48550/arxiv.2411.05383
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2411_05383</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2411_05383</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2411_053833</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjE01DMwNbYw5mSwDskvTyxKKVbwyS_XDUotzi8tSk5V8Egsyk0rzVHwTc1NVXBJLUlNLsnMz1MozyzJUPDx9VVwTE_NKynmYWBNS8wpTuWF0twM8m6uIc4eumBr4guKMnMTiyrjQdbFg60zJqwCAH8FNCE</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><source>arXiv.org</source><creator>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</creator><creatorcontrib>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</creatorcontrib><description>The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an agency-driven framework for low-resource harmful meme detection, employing both outward and inward analysis with few-shot annotated samples. Inspired by the powerful capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first retrieve relative memes with annotations to leverage label information as auxiliary signals for the LMM agent. Then, we elicit knowledge-revising behavior within the LMM agent to derive well-generalized insights into meme harmfulness. By combining these strategies, our approach enables dialectical reasoning over intricate and implicit harm-indicative patterns. Extensive experiments conducted on three meme datasets demonstrate that our proposed approach achieves superior performance than state-of-the-art methods on the low-resource harmful meme detection task.</description><identifier>DOI: 10.48550/arxiv.2411.05383</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2024-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2411.05383$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2411.05383$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Huang, Jianzhao</creatorcontrib><creatorcontrib>Lin, Hongzhan</creatorcontrib><creatorcontrib>Liu, Ziyan</creatorcontrib><creatorcontrib>Luo, Ziyang</creatorcontrib><creatorcontrib>Chen, Guang</creatorcontrib><creatorcontrib>Ma, Jing</creatorcontrib><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><description>The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an agency-driven framework for low-resource harmful meme detection, employing both outward and inward analysis with few-shot annotated samples. Inspired by the powerful capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first retrieve relative memes with annotations to leverage label information as auxiliary signals for the LMM agent. Then, we elicit knowledge-revising behavior within the LMM agent to derive well-generalized insights into meme harmfulness. By combining these strategies, our approach enables dialectical reasoning over intricate and implicit harm-indicative patterns. Extensive experiments conducted on three meme datasets demonstrate that our proposed approach achieves superior performance than state-of-the-art methods on the low-resource harmful meme detection task.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjE01DMwNbYw5mSwDskvTyxKKVbwyS_XDUotzi8tSk5V8Egsyk0rzVHwTc1NVXBJLUlNLsnMz1MozyzJUPDx9VVwTE_NKynmYWBNS8wpTuWF0twM8m6uIc4eumBr4guKMnMTiyrjQdbFg60zJqwCAH8FNCE</recordid><startdate>20241108</startdate><enddate>20241108</enddate><creator>Huang, Jianzhao</creator><creator>Lin, Hongzhan</creator><creator>Liu, Ziyan</creator><creator>Luo, Ziyang</creator><creator>Chen, Guang</creator><creator>Ma, Jing</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20241108</creationdate><title>Towards Low-Resource Harmful Meme Detection with LMM Agents</title><author>Huang, Jianzhao ; Lin, Hongzhan ; Liu, Ziyan ; Luo, Ziyang ; Chen, Guang ; Ma, Jing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2411_053833</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Huang, Jianzhao</creatorcontrib><creatorcontrib>Lin, Hongzhan</creatorcontrib><creatorcontrib>Liu, Ziyan</creatorcontrib><creatorcontrib>Luo, Ziyang</creatorcontrib><creatorcontrib>Chen, Guang</creatorcontrib><creatorcontrib>Ma, Jing</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Huang, Jianzhao</au><au>Lin, Hongzhan</au><au>Liu, Ziyan</au><au>Luo, Ziyang</au><au>Chen, Guang</au><au>Ma, Jing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Towards Low-Resource Harmful Meme Detection with LMM Agents</atitle><date>2024-11-08</date><risdate>2024</risdate><abstract>The proliferation of Internet memes in the age of social media necessitates effective identification of harmful ones. Due to the dynamic nature of memes, existing data-driven models may struggle in low-resource scenarios where only a few labeled examples are available. In this paper, we propose an agency-driven framework for low-resource harmful meme detection, employing both outward and inward analysis with few-shot annotated samples. Inspired by the powerful capacity of Large Multimodal Models (LMMs) on multimodal reasoning, we first retrieve relative memes with annotations to leverage label information as auxiliary signals for the LMM agent. Then, we elicit knowledge-revising behavior within the LMM agent to derive well-generalized insights into meme harmfulness. By combining these strategies, our approach enables dialectical reasoning over intricate and implicit harm-indicative patterns. Extensive experiments conducted on three meme datasets demonstrate that our proposed approach achieves superior performance than state-of-the-art methods on the low-resource harmful meme detection task.</abstract><doi>10.48550/arxiv.2411.05383</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2411.05383
ispartof
issn
language eng
recordid cdi_arxiv_primary_2411_05383
source arXiv.org
subjects Computer Science - Computation and Language
title Towards Low-Resource Harmful Meme Detection with LMM Agents
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T02%3A01%3A16IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Towards%20Low-Resource%20Harmful%20Meme%20Detection%20with%20LMM%20Agents&rft.au=Huang,%20Jianzhao&rft.date=2024-11-08&rft_id=info:doi/10.48550/arxiv.2411.05383&rft_dat=%3Carxiv_GOX%3E2411_05383%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true