A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis

Multi-modal aspect-based sentiment analysis (MABSA) has recently attracted increasing attention. The span-based extraction methods, such as FSUIE, demonstrate strong performance in sentiment analysis due to their joint modeling of input sequences and target labels. However, previous methods still ha...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2023-12
Hauptverfasser: Peng, Tianshuo, Li, Zuchao, Wang, Ping, Zhang, Lefei, Zhao, Hai
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Peng, Tianshuo
Li, Zuchao
Wang, Ping
Zhang, Lefei
Zhao, Hai
description Multi-modal aspect-based sentiment analysis (MABSA) has recently attracted increasing attention. The span-based extraction methods, such as FSUIE, demonstrate strong performance in sentiment analysis due to their joint modeling of input sequences and target labels. However, previous methods still have certain limitations: (i) They ignore the difference in the focus of visual information between different analysis targets (aspect or sentiment). (ii) Combining features from uni-modal encoders directly may not be sufficient to eliminate the modal gap and can cause difficulties in capturing the image-text pairwise relevance. (iii) Existing span-based methods for MABSA ignore the pairwise relevance of target span boundaries. To tackle these limitations, we propose a novel framework called DQPSA for multi-modal sentiment analysis. Specifically, our model contains a Prompt as Dual Query (PDQ) module that uses the prompt as both a visual query and a language query to extract prompt-aware visual information and strengthen the pairwise relevance between visual information and the analysis target. Additionally, we introduce an Energy-based Pairwise Expert (EPE) module that models the boundaries pairing of the analysis target from the perspective of an Energy-based Model. This expert predicts aspect or sentiment span based on pairwise stability. Experiments on three widely used benchmarks demonstrate that DQPSA outperforms previous approaches and achieves a new state-of-the-art performance.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2901611184</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2901611184</sourcerecordid><originalsourceid>FETCH-proquest_journals_29016111843</originalsourceid><addsrcrecordid>eNqNyk8LgjAcxvERBEn5HgadB9v8kx0tjC52qbssnTWZm-03A999Er2ALs8XHj4LFPAoYiSLOV-hEKCjlPJ0x5MkCtAtxxf7lhoXRrrHhO8CZINL28xXKeunMAp63FqHy1F7RXrbCI1zGGTtyeGLr9J41c-DcyP0BAo2aNkKDTL8dY22p-J2PJPB2dcowVedHd2MoeJ7ylLGWBZH_6kP6Uw_gg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2901611184</pqid></control><display><type>article</type><title>A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis</title><source>Free E- Journals</source><creator>Peng, Tianshuo ; Li, Zuchao ; Wang, Ping ; Zhang, Lefei ; Zhao, Hai</creator><creatorcontrib>Peng, Tianshuo ; Li, Zuchao ; Wang, Ping ; Zhang, Lefei ; Zhao, Hai</creatorcontrib><description>Multi-modal aspect-based sentiment analysis (MABSA) has recently attracted increasing attention. The span-based extraction methods, such as FSUIE, demonstrate strong performance in sentiment analysis due to their joint modeling of input sequences and target labels. However, previous methods still have certain limitations: (i) They ignore the difference in the focus of visual information between different analysis targets (aspect or sentiment). (ii) Combining features from uni-modal encoders directly may not be sufficient to eliminate the modal gap and can cause difficulties in capturing the image-text pairwise relevance. (iii) Existing span-based methods for MABSA ignore the pairwise relevance of target span boundaries. To tackle these limitations, we propose a novel framework called DQPSA for multi-modal sentiment analysis. Specifically, our model contains a Prompt as Dual Query (PDQ) module that uses the prompt as both a visual query and a language query to extract prompt-aware visual information and strengthen the pairwise relevance between visual information and the analysis target. Additionally, we introduce an Energy-based Pairwise Expert (EPE) module that models the boundaries pairing of the analysis target from the perspective of an Energy-based Model. This expert predicts aspect or sentiment span based on pairwise stability. Experiments on three widely used benchmarks demonstrate that DQPSA outperforms previous approaches and achieves a new state-of-the-art performance.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Boundaries ; Data mining ; Modules ; Queries ; Query languages ; Sentiment analysis</subject><ispartof>arXiv.org, 2023-12</ispartof><rights>2023. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Peng, Tianshuo</creatorcontrib><creatorcontrib>Li, Zuchao</creatorcontrib><creatorcontrib>Wang, Ping</creatorcontrib><creatorcontrib>Zhang, Lefei</creatorcontrib><creatorcontrib>Zhao, Hai</creatorcontrib><title>A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis</title><title>arXiv.org</title><description>Multi-modal aspect-based sentiment analysis (MABSA) has recently attracted increasing attention. The span-based extraction methods, such as FSUIE, demonstrate strong performance in sentiment analysis due to their joint modeling of input sequences and target labels. However, previous methods still have certain limitations: (i) They ignore the difference in the focus of visual information between different analysis targets (aspect or sentiment). (ii) Combining features from uni-modal encoders directly may not be sufficient to eliminate the modal gap and can cause difficulties in capturing the image-text pairwise relevance. (iii) Existing span-based methods for MABSA ignore the pairwise relevance of target span boundaries. To tackle these limitations, we propose a novel framework called DQPSA for multi-modal sentiment analysis. Specifically, our model contains a Prompt as Dual Query (PDQ) module that uses the prompt as both a visual query and a language query to extract prompt-aware visual information and strengthen the pairwise relevance between visual information and the analysis target. Additionally, we introduce an Energy-based Pairwise Expert (EPE) module that models the boundaries pairing of the analysis target from the perspective of an Energy-based Model. This expert predicts aspect or sentiment span based on pairwise stability. Experiments on three widely used benchmarks demonstrate that DQPSA outperforms previous approaches and achieves a new state-of-the-art performance.</description><subject>Boundaries</subject><subject>Data mining</subject><subject>Modules</subject><subject>Queries</subject><subject>Query languages</subject><subject>Sentiment analysis</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNyk8LgjAcxvERBEn5HgadB9v8kx0tjC52qbssnTWZm-03A999Er2ALs8XHj4LFPAoYiSLOV-hEKCjlPJ0x5MkCtAtxxf7lhoXRrrHhO8CZINL28xXKeunMAp63FqHy1F7RXrbCI1zGGTtyeGLr9J41c-DcyP0BAo2aNkKDTL8dY22p-J2PJPB2dcowVedHd2MoeJ7ylLGWBZH_6kP6Uw_gg</recordid><startdate>20231215</startdate><enddate>20231215</enddate><creator>Peng, Tianshuo</creator><creator>Li, Zuchao</creator><creator>Wang, Ping</creator><creator>Zhang, Lefei</creator><creator>Zhao, Hai</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20231215</creationdate><title>A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis</title><author>Peng, Tianshuo ; Li, Zuchao ; Wang, Ping ; Zhang, Lefei ; Zhao, Hai</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_29016111843</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Boundaries</topic><topic>Data mining</topic><topic>Modules</topic><topic>Queries</topic><topic>Query languages</topic><topic>Sentiment analysis</topic><toplevel>online_resources</toplevel><creatorcontrib>Peng, Tianshuo</creatorcontrib><creatorcontrib>Li, Zuchao</creatorcontrib><creatorcontrib>Wang, Ping</creatorcontrib><creatorcontrib>Zhang, Lefei</creatorcontrib><creatorcontrib>Zhao, Hai</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni)</collection><collection>ProQuest Central</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Peng, Tianshuo</au><au>Li, Zuchao</au><au>Wang, Ping</au><au>Zhang, Lefei</au><au>Zhao, Hai</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis</atitle><jtitle>arXiv.org</jtitle><date>2023-12-15</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>Multi-modal aspect-based sentiment analysis (MABSA) has recently attracted increasing attention. The span-based extraction methods, such as FSUIE, demonstrate strong performance in sentiment analysis due to their joint modeling of input sequences and target labels. However, previous methods still have certain limitations: (i) They ignore the difference in the focus of visual information between different analysis targets (aspect or sentiment). (ii) Combining features from uni-modal encoders directly may not be sufficient to eliminate the modal gap and can cause difficulties in capturing the image-text pairwise relevance. (iii) Existing span-based methods for MABSA ignore the pairwise relevance of target span boundaries. To tackle these limitations, we propose a novel framework called DQPSA for multi-modal sentiment analysis. Specifically, our model contains a Prompt as Dual Query (PDQ) module that uses the prompt as both a visual query and a language query to extract prompt-aware visual information and strengthen the pairwise relevance between visual information and the analysis target. Additionally, we introduce an Energy-based Pairwise Expert (EPE) module that models the boundaries pairing of the analysis target from the perspective of an Energy-based Model. This expert predicts aspect or sentiment span based on pairwise stability. Experiments on three widely used benchmarks demonstrate that DQPSA outperforms previous approaches and achieves a new state-of-the-art performance.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2023-12
issn 2331-8422
language eng
recordid cdi_proquest_journals_2901611184
source Free E- Journals
subjects Boundaries
Data mining
Modules
Queries
Query languages
Sentiment analysis
title A Novel Energy based Model Mechanism for Multi-modal Aspect-Based Sentiment Analysis
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T06%3A01%3A39IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20Novel%20Energy%20based%20Model%20Mechanism%20for%20Multi-modal%20Aspect-Based%20Sentiment%20Analysis&rft.jtitle=arXiv.org&rft.au=Peng,%20Tianshuo&rft.date=2023-12-15&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2901611184%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2901611184&rft_id=info:pmid/&rfr_iscdi=true