Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection

Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-b...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Mastromattei, Michele, Zanzotto, Fabio Massimo
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Mastromattei, Michele
Zanzotto, Fabio Massimo
description Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.
doi_str_mv 10.48550/arxiv.2406.02338
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_02338</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_02338</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2406_023383</originalsourceid><addsrcrecordid>eNqFjs8KgkAQxvfSIaoH6NS8QGapIV0rUTAIkq4y2CgLuhuzavn2bdK903zM94efEMut6_hhELgb5LfsnZ3v7h1353nhVPSpVFUnTSsLiKwkfrJULUgFGaMypeaGGC76QbU5QKxfkKJtYEVwR5bYSq0gUWXdkSrIwBUZG2pt50Y1FaNtxxLWaoCTNcbXXExKrA0tfncmVtE5O8brkTC3DA3ykH9J85HU-5_4AGM_SyI</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><source>arXiv.org</source><creator>Mastromattei, Michele ; Zanzotto, Fabio Massimo</creator><creatorcontrib>Mastromattei, Michele ; Zanzotto, Fabio Massimo</creatorcontrib><description>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.</description><identifier>DOI: 10.48550/arxiv.2406.02338</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2024-06</creationdate><rights>http://creativecommons.org/publicdomain/zero/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.02338$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.02338$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mastromattei, Michele</creatorcontrib><creatorcontrib>Zanzotto, Fabio Massimo</creatorcontrib><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><description>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFjs8KgkAQxvfSIaoH6NS8QGapIV0rUTAIkq4y2CgLuhuzavn2bdK903zM94efEMut6_hhELgb5LfsnZ3v7h1353nhVPSpVFUnTSsLiKwkfrJULUgFGaMypeaGGC76QbU5QKxfkKJtYEVwR5bYSq0gUWXdkSrIwBUZG2pt50Y1FaNtxxLWaoCTNcbXXExKrA0tfncmVtE5O8brkTC3DA3ykH9J85HU-5_4AGM_SyI</recordid><startdate>20240604</startdate><enddate>20240604</enddate><creator>Mastromattei, Michele</creator><creator>Zanzotto, Fabio Massimo</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240604</creationdate><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><author>Mastromattei, Michele ; Zanzotto, Fabio Massimo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2406_023383</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Mastromattei, Michele</creatorcontrib><creatorcontrib>Zanzotto, Fabio Massimo</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mastromattei, Michele</au><au>Zanzotto, Fabio Massimo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</atitle><date>2024-06-04</date><risdate>2024</risdate><abstract>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-based models for irony detection. To conduct our study, we used the EPIC corpus to extract five diverse English variation-specific datasets and applied the KEN pruning algorithm on five different architectures. Our results reveal several similarities between optimal subnetworks, which provide insights into the linguistic variations that share strong resemblances and those that exhibit greater dissimilarities. We discovered that optimal subnetworks across models share at least 60% of their parameters, emphasizing the significance of parameter values in capturing and interpreting linguistic variations. This study highlights the inherent structural similarities between models trained on different variants of the same language and also the critical role of parameter values in capturing these nuances.</abstract><doi>10.48550/arxiv.2406.02338</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2406.02338
ispartof
issn
language eng
recordid cdi_arxiv_primary_2406_02338
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
title Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T04%3A27%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Linguistic%20Fingerprint%20in%20Transformer%20Models:%20How%20Language%20Variation%20Influences%20Parameter%20Selection%20in%20Irony%20Detection&rft.au=Mastromattei,%20Michele&rft.date=2024-06-04&rft_id=info:doi/10.48550/arxiv.2406.02338&rft_dat=%3Carxiv_GOX%3E2406_02338%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true