Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective
Prior studies on the emergence in large models have primarily focused on how the functional capabilities of large language models (LLMs) scale with model size. Our research, however, transcends this traditional paradigm, aiming to deepen our understanding of the emergence within LLMs by placing a sp...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Xiao, Xiongye Zhou, Chenyu Ping, Heng Cao, Defu Li, Yaxing Zhou, Yizhuo Li, Shixuan Bogdan, Paul |
description | Prior studies on the emergence in large models have primarily focused on how
the functional capabilities of large language models (LLMs) scale with model
size. Our research, however, transcends this traditional paradigm, aiming to
deepen our understanding of the emergence within LLMs by placing a special
emphasis not just on the model size but more significantly on the complex
behavior of neuron interactions during the training process. By introducing the
concepts of "self-organization" and "multifractal analysis," we explore how
neuron interactions dynamically evolve during training, leading to "emergence,"
mirroring the phenomenon in natural systems where simple micro-level
interactions give rise to complex macro-level behaviors. To quantitatively
analyze the continuously evolving interactions among neurons in large models
during training, we propose the Neuron-based Multifractal Analysis (NeuroMFA).
Utilizing NeuroMFA, we conduct a comprehensive examination of the emergent
behavior in LLMs through the lens of both model size and training process,
paving new avenues for research into the emergence in large models. |
doi_str_mv | 10.48550/arxiv.2402.09099 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2402_09099</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2402_09099</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-61fd04f0b701a283796cf4142cb690425850f616c1997bfbff4f947309afd69c3</originalsourceid><addsrcrecordid>eNotz7FOwzAUhWEvDKjwAEzcF0i4Thwnl62qWqiUAkP3yHHsYilxIjut2reHFqazHP3Sx9gTx1RURYEvKpzdKc0EZikSEt2zZn2e-jE4f4APcwyjh62fTVB6dqOPoHwH68GEg_HagPNQ17v4CpswDjB_G9gd-9nZ6131sPSqv0QX4cuEOJnfxMk8sDur-mge_3fB9pv1fvWe1J9v29WyTpQsKZHcdigstiVylVV5SVJbwUWmW0kosqIq0EouNScqW9taKyyJMkdStpOk8wV7_svehM0U3KDCpblKm5s0_wFdZk6p</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective</title><source>arXiv.org</source><creator>Xiao, Xiongye ; Zhou, Chenyu ; Ping, Heng ; Cao, Defu ; Li, Yaxing ; Zhou, Yizhuo ; Li, Shixuan ; Bogdan, Paul</creator><creatorcontrib>Xiao, Xiongye ; Zhou, Chenyu ; Ping, Heng ; Cao, Defu ; Li, Yaxing ; Zhou, Yizhuo ; Li, Shixuan ; Bogdan, Paul</creatorcontrib><description>Prior studies on the emergence in large models have primarily focused on how
the functional capabilities of large language models (LLMs) scale with model
size. Our research, however, transcends this traditional paradigm, aiming to
deepen our understanding of the emergence within LLMs by placing a special
emphasis not just on the model size but more significantly on the complex
behavior of neuron interactions during the training process. By introducing the
concepts of "self-organization" and "multifractal analysis," we explore how
neuron interactions dynamically evolve during training, leading to "emergence,"
mirroring the phenomenon in natural systems where simple micro-level
interactions give rise to complex macro-level behaviors. To quantitatively
analyze the continuously evolving interactions among neurons in large models
during training, we propose the Neuron-based Multifractal Analysis (NeuroMFA).
Utilizing NeuroMFA, we conduct a comprehensive examination of the emergent
behavior in LLMs through the lens of both model size and training process,
paving new avenues for research into the emergence in large models.</description><identifier>DOI: 10.48550/arxiv.2402.09099</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence</subject><creationdate>2024-02</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2402.09099$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2402.09099$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Xiao, Xiongye</creatorcontrib><creatorcontrib>Zhou, Chenyu</creatorcontrib><creatorcontrib>Ping, Heng</creatorcontrib><creatorcontrib>Cao, Defu</creatorcontrib><creatorcontrib>Li, Yaxing</creatorcontrib><creatorcontrib>Zhou, Yizhuo</creatorcontrib><creatorcontrib>Li, Shixuan</creatorcontrib><creatorcontrib>Bogdan, Paul</creatorcontrib><title>Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective</title><description>Prior studies on the emergence in large models have primarily focused on how
the functional capabilities of large language models (LLMs) scale with model
size. Our research, however, transcends this traditional paradigm, aiming to
deepen our understanding of the emergence within LLMs by placing a special
emphasis not just on the model size but more significantly on the complex
behavior of neuron interactions during the training process. By introducing the
concepts of "self-organization" and "multifractal analysis," we explore how
neuron interactions dynamically evolve during training, leading to "emergence,"
mirroring the phenomenon in natural systems where simple micro-level
interactions give rise to complex macro-level behaviors. To quantitatively
analyze the continuously evolving interactions among neurons in large models
during training, we propose the Neuron-based Multifractal Analysis (NeuroMFA).
Utilizing NeuroMFA, we conduct a comprehensive examination of the emergent
behavior in LLMs through the lens of both model size and training process,
paving new avenues for research into the emergence in large models.</description><subject>Computer Science - Artificial Intelligence</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz7FOwzAUhWEvDKjwAEzcF0i4Thwnl62qWqiUAkP3yHHsYilxIjut2reHFqazHP3Sx9gTx1RURYEvKpzdKc0EZikSEt2zZn2e-jE4f4APcwyjh62fTVB6dqOPoHwH68GEg_HagPNQ17v4CpswDjB_G9gd-9nZ6131sPSqv0QX4cuEOJnfxMk8sDur-mge_3fB9pv1fvWe1J9v29WyTpQsKZHcdigstiVylVV5SVJbwUWmW0kosqIq0EouNScqW9taKyyJMkdStpOk8wV7_svehM0U3KDCpblKm5s0_wFdZk6p</recordid><startdate>20240214</startdate><enddate>20240214</enddate><creator>Xiao, Xiongye</creator><creator>Zhou, Chenyu</creator><creator>Ping, Heng</creator><creator>Cao, Defu</creator><creator>Li, Yaxing</creator><creator>Zhou, Yizhuo</creator><creator>Li, Shixuan</creator><creator>Bogdan, Paul</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240214</creationdate><title>Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective</title><author>Xiao, Xiongye ; Zhou, Chenyu ; Ping, Heng ; Cao, Defu ; Li, Yaxing ; Zhou, Yizhuo ; Li, Shixuan ; Bogdan, Paul</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-61fd04f0b701a283796cf4142cb690425850f616c1997bfbff4f947309afd69c3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><toplevel>online_resources</toplevel><creatorcontrib>Xiao, Xiongye</creatorcontrib><creatorcontrib>Zhou, Chenyu</creatorcontrib><creatorcontrib>Ping, Heng</creatorcontrib><creatorcontrib>Cao, Defu</creatorcontrib><creatorcontrib>Li, Yaxing</creatorcontrib><creatorcontrib>Zhou, Yizhuo</creatorcontrib><creatorcontrib>Li, Shixuan</creatorcontrib><creatorcontrib>Bogdan, Paul</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Xiao, Xiongye</au><au>Zhou, Chenyu</au><au>Ping, Heng</au><au>Cao, Defu</au><au>Li, Yaxing</au><au>Zhou, Yizhuo</au><au>Li, Shixuan</au><au>Bogdan, Paul</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective</atitle><date>2024-02-14</date><risdate>2024</risdate><abstract>Prior studies on the emergence in large models have primarily focused on how
the functional capabilities of large language models (LLMs) scale with model
size. Our research, however, transcends this traditional paradigm, aiming to
deepen our understanding of the emergence within LLMs by placing a special
emphasis not just on the model size but more significantly on the complex
behavior of neuron interactions during the training process. By introducing the
concepts of "self-organization" and "multifractal analysis," we explore how
neuron interactions dynamically evolve during training, leading to "emergence,"
mirroring the phenomenon in natural systems where simple micro-level
interactions give rise to complex macro-level behaviors. To quantitatively
analyze the continuously evolving interactions among neurons in large models
during training, we propose the Neuron-based Multifractal Analysis (NeuroMFA).
Utilizing NeuroMFA, we conduct a comprehensive examination of the emergent
behavior in LLMs through the lens of both model size and training process,
paving new avenues for research into the emergence in large models.</abstract><doi>10.48550/arxiv.2402.09099</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2402.09099 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2402_09099 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence |
title | Exploring Neuron Interactions and Emergence in LLMs: From the Multifractal Analysis Perspective |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-27T22%3A06%3A34IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Exploring%20Neuron%20Interactions%20and%20Emergence%20in%20LLMs:%20From%20the%20Multifractal%20Analysis%20Perspective&rft.au=Xiao,%20Xiongye&rft.date=2024-02-14&rft_id=info:doi/10.48550/arxiv.2402.09099&rft_dat=%3Carxiv_GOX%3E2402_09099%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |