Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection
Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP (NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment analysis and transformer model architectures. We aim to investigate how different English variations impact transformer-b...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Mastromattei, Michele Zanzotto, Fabio Massimo |
description | Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP
(NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment
analysis and transformer model architectures. We aim to investigate how
different English variations impact transformer-based models for irony
detection. To conduct our study, we used the EPIC corpus to extract five
diverse English variation-specific datasets and applied the KEN pruning
algorithm on five different architectures. Our results reveal several
similarities between optimal subnetworks, which provide insights into the
linguistic variations that share strong resemblances and those that exhibit
greater dissimilarities. We discovered that optimal subnetworks across models
share at least 60% of their parameters, emphasizing the significance of
parameter values in capturing and interpreting linguistic variations. This
study highlights the inherent structural similarities between models trained on
different variants of the same language and also the critical role of parameter
values in capturing these nuances. |
doi_str_mv | 10.48550/arxiv.2406.02338 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_02338</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_02338</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2406_023383</originalsourceid><addsrcrecordid>eNqFjs8KgkAQxvfSIaoH6NS8QGapIV0rUTAIkq4y2CgLuhuzavn2bdK903zM94efEMut6_hhELgb5LfsnZ3v7h1353nhVPSpVFUnTSsLiKwkfrJULUgFGaMypeaGGC76QbU5QKxfkKJtYEVwR5bYSq0gUWXdkSrIwBUZG2pt50Y1FaNtxxLWaoCTNcbXXExKrA0tfncmVtE5O8brkTC3DA3ykH9J85HU-5_4AGM_SyI</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><source>arXiv.org</source><creator>Mastromattei, Michele ; Zanzotto, Fabio Massimo</creator><creatorcontrib>Mastromattei, Michele ; Zanzotto, Fabio Massimo</creatorcontrib><description>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP
(NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment
analysis and transformer model architectures. We aim to investigate how
different English variations impact transformer-based models for irony
detection. To conduct our study, we used the EPIC corpus to extract five
diverse English variation-specific datasets and applied the KEN pruning
algorithm on five different architectures. Our results reveal several
similarities between optimal subnetworks, which provide insights into the
linguistic variations that share strong resemblances and those that exhibit
greater dissimilarities. We discovered that optimal subnetworks across models
share at least 60% of their parameters, emphasizing the significance of
parameter values in capturing and interpreting linguistic variations. This
study highlights the inherent structural similarities between models trained on
different variants of the same language and also the critical role of parameter
values in capturing these nuances.</description><identifier>DOI: 10.48550/arxiv.2406.02338</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2024-06</creationdate><rights>http://creativecommons.org/publicdomain/zero/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.02338$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.02338$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mastromattei, Michele</creatorcontrib><creatorcontrib>Zanzotto, Fabio Massimo</creatorcontrib><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><description>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP
(NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment
analysis and transformer model architectures. We aim to investigate how
different English variations impact transformer-based models for irony
detection. To conduct our study, we used the EPIC corpus to extract five
diverse English variation-specific datasets and applied the KEN pruning
algorithm on five different architectures. Our results reveal several
similarities between optimal subnetworks, which provide insights into the
linguistic variations that share strong resemblances and those that exhibit
greater dissimilarities. We discovered that optimal subnetworks across models
share at least 60% of their parameters, emphasizing the significance of
parameter values in capturing and interpreting linguistic variations. This
study highlights the inherent structural similarities between models trained on
different variants of the same language and also the critical role of parameter
values in capturing these nuances.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFjs8KgkAQxvfSIaoH6NS8QGapIV0rUTAIkq4y2CgLuhuzavn2bdK903zM94efEMut6_hhELgb5LfsnZ3v7h1353nhVPSpVFUnTSsLiKwkfrJULUgFGaMypeaGGC76QbU5QKxfkKJtYEVwR5bYSq0gUWXdkSrIwBUZG2pt50Y1FaNtxxLWaoCTNcbXXExKrA0tfncmVtE5O8brkTC3DA3ykH9J85HU-5_4AGM_SyI</recordid><startdate>20240604</startdate><enddate>20240604</enddate><creator>Mastromattei, Michele</creator><creator>Zanzotto, Fabio Massimo</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240604</creationdate><title>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</title><author>Mastromattei, Michele ; Zanzotto, Fabio Massimo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2406_023383</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Mastromattei, Michele</creatorcontrib><creatorcontrib>Zanzotto, Fabio Massimo</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mastromattei, Michele</au><au>Zanzotto, Fabio Massimo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection</atitle><date>2024-06-04</date><risdate>2024</risdate><abstract>Proceedings of the 3rd Workshop on Perspectivist Approaches to NLP
(NLPerspectives) @ LREC-COLING 2024 This paper explores the correlation between linguistic diversity, sentiment
analysis and transformer model architectures. We aim to investigate how
different English variations impact transformer-based models for irony
detection. To conduct our study, we used the EPIC corpus to extract five
diverse English variation-specific datasets and applied the KEN pruning
algorithm on five different architectures. Our results reveal several
similarities between optimal subnetworks, which provide insights into the
linguistic variations that share strong resemblances and those that exhibit
greater dissimilarities. We discovered that optimal subnetworks across models
share at least 60% of their parameters, emphasizing the significance of
parameter values in capturing and interpreting linguistic variations. This
study highlights the inherent structural similarities between models trained on
different variants of the same language and also the critical role of parameter
values in capturing these nuances.</abstract><doi>10.48550/arxiv.2406.02338</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2406.02338 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2406_02338 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Computation and Language |
title | Linguistic Fingerprint in Transformer Models: How Language Variation Influences Parameter Selection in Irony Detection |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T04%3A27%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Linguistic%20Fingerprint%20in%20Transformer%20Models:%20How%20Language%20Variation%20Influences%20Parameter%20Selection%20in%20Irony%20Detection&rft.au=Mastromattei,%20Michele&rft.date=2024-06-04&rft_id=info:doi/10.48550/arxiv.2406.02338&rft_dat=%3Carxiv_GOX%3E2406_02338%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |