Implicit variance regularization in non-contrastive SSL
Non-contrastive SSL methods like BYOL and SimSiam rely on asymmetric predictor networks to avoid representational collapse without negative samples. Yet, how predictor networks facilitate stable learning is not fully understood. While previous theoretical analyses assumed Euclidean losses, most prac...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Halvagal, Manu Srinath Laborieux, Axel Zenke, Friedemann |
description | Non-contrastive SSL methods like BYOL and SimSiam rely on asymmetric
predictor networks to avoid representational collapse without negative samples.
Yet, how predictor networks facilitate stable learning is not fully understood.
While previous theoretical analyses assumed Euclidean losses, most practical
implementations rely on cosine similarity. To gain further theoretical insight
into non-contrastive SSL, we analytically study learning dynamics in
conjunction with Euclidean and cosine similarity in the eigenspace of
closed-form linear predictor networks. We show that both avoid collapse through
implicit variance regularization albeit through different dynamical mechanisms.
Moreover, we find that the eigenvalues act as effective learning rate
multipliers and propose a family of isotropic loss functions (IsoLoss) that
equalize convergence rates across eigenmodes. Empirically, IsoLoss speeds up
the initial learning dynamics and increases robustness, thereby allowing us to
dispense with the EMA target network typically used with non-contrastive
methods. Our analysis sheds light on the variance regularization mechanisms of
non-contrastive SSL and lays the theoretical grounds for crafting novel loss
functions that shape the learning dynamics of the predictor's spectrum. |
doi_str_mv | 10.48550/arxiv.2212.04858 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2212_04858</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2212_04858</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-bc9ab43a830afd6d3c19642ee6df5ed20ee1156e258063cae3b47679a02c77423</originalsourceid><addsrcrecordid>eNotj8tuwjAURL1hUUE_oKv6BxL8trNECFqkSCxgH904N5Wl4CATIuDry2s1mlkczSHki7NcOa3ZHNIljLkQXOTsvrgPYjeHYxd8GOgIKUD0SBP-nbt7ucEQ-khDpLGPme_jkOA0hBHpblfOyKSF7oSf75yS_Xq1X_5m5fZns1yUGRjrstoXUCsJTjJoG9NIzwujBKJpWo2NYIica4NCO2akB5S1ssYWwIS3Vgk5Jd8v7PN5dUzhAOlaPQyqp4H8B7A4QPM</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Implicit variance regularization in non-contrastive SSL</title><source>arXiv.org</source><creator>Halvagal, Manu Srinath ; Laborieux, Axel ; Zenke, Friedemann</creator><creatorcontrib>Halvagal, Manu Srinath ; Laborieux, Axel ; Zenke, Friedemann</creatorcontrib><description>Non-contrastive SSL methods like BYOL and SimSiam rely on asymmetric
predictor networks to avoid representational collapse without negative samples.
Yet, how predictor networks facilitate stable learning is not fully understood.
While previous theoretical analyses assumed Euclidean losses, most practical
implementations rely on cosine similarity. To gain further theoretical insight
into non-contrastive SSL, we analytically study learning dynamics in
conjunction with Euclidean and cosine similarity in the eigenspace of
closed-form linear predictor networks. We show that both avoid collapse through
implicit variance regularization albeit through different dynamical mechanisms.
Moreover, we find that the eigenvalues act as effective learning rate
multipliers and propose a family of isotropic loss functions (IsoLoss) that
equalize convergence rates across eigenmodes. Empirically, IsoLoss speeds up
the initial learning dynamics and increases robustness, thereby allowing us to
dispense with the EMA target network typically used with non-contrastive
methods. Our analysis sheds light on the variance regularization mechanisms of
non-contrastive SSL and lays the theoretical grounds for crafting novel loss
functions that shape the learning dynamics of the predictor's spectrum.</description><identifier>DOI: 10.48550/arxiv.2212.04858</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning ; Computer Science - Neural and Evolutionary Computing</subject><creationdate>2022-12</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2212.04858$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2212.04858$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Halvagal, Manu Srinath</creatorcontrib><creatorcontrib>Laborieux, Axel</creatorcontrib><creatorcontrib>Zenke, Friedemann</creatorcontrib><title>Implicit variance regularization in non-contrastive SSL</title><description>Non-contrastive SSL methods like BYOL and SimSiam rely on asymmetric
predictor networks to avoid representational collapse without negative samples.
Yet, how predictor networks facilitate stable learning is not fully understood.
While previous theoretical analyses assumed Euclidean losses, most practical
implementations rely on cosine similarity. To gain further theoretical insight
into non-contrastive SSL, we analytically study learning dynamics in
conjunction with Euclidean and cosine similarity in the eigenspace of
closed-form linear predictor networks. We show that both avoid collapse through
implicit variance regularization albeit through different dynamical mechanisms.
Moreover, we find that the eigenvalues act as effective learning rate
multipliers and propose a family of isotropic loss functions (IsoLoss) that
equalize convergence rates across eigenmodes. Empirically, IsoLoss speeds up
the initial learning dynamics and increases robustness, thereby allowing us to
dispense with the EMA target network typically used with non-contrastive
methods. Our analysis sheds light on the variance regularization mechanisms of
non-contrastive SSL and lays the theoretical grounds for crafting novel loss
functions that shape the learning dynamics of the predictor's spectrum.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Computer Science - Neural and Evolutionary Computing</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tuwjAURL1hUUE_oKv6BxL8trNECFqkSCxgH904N5Wl4CATIuDry2s1mlkczSHki7NcOa3ZHNIljLkQXOTsvrgPYjeHYxd8GOgIKUD0SBP-nbt7ucEQ-khDpLGPme_jkOA0hBHpblfOyKSF7oSf75yS_Xq1X_5m5fZns1yUGRjrstoXUCsJTjJoG9NIzwujBKJpWo2NYIica4NCO2akB5S1ssYWwIS3Vgk5Jd8v7PN5dUzhAOlaPQyqp4H8B7A4QPM</recordid><startdate>20221209</startdate><enddate>20221209</enddate><creator>Halvagal, Manu Srinath</creator><creator>Laborieux, Axel</creator><creator>Zenke, Friedemann</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221209</creationdate><title>Implicit variance regularization in non-contrastive SSL</title><author>Halvagal, Manu Srinath ; Laborieux, Axel ; Zenke, Friedemann</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-bc9ab43a830afd6d3c19642ee6df5ed20ee1156e258063cae3b47679a02c77423</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Computer Science - Neural and Evolutionary Computing</topic><toplevel>online_resources</toplevel><creatorcontrib>Halvagal, Manu Srinath</creatorcontrib><creatorcontrib>Laborieux, Axel</creatorcontrib><creatorcontrib>Zenke, Friedemann</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Halvagal, Manu Srinath</au><au>Laborieux, Axel</au><au>Zenke, Friedemann</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Implicit variance regularization in non-contrastive SSL</atitle><date>2022-12-09</date><risdate>2022</risdate><abstract>Non-contrastive SSL methods like BYOL and SimSiam rely on asymmetric
predictor networks to avoid representational collapse without negative samples.
Yet, how predictor networks facilitate stable learning is not fully understood.
While previous theoretical analyses assumed Euclidean losses, most practical
implementations rely on cosine similarity. To gain further theoretical insight
into non-contrastive SSL, we analytically study learning dynamics in
conjunction with Euclidean and cosine similarity in the eigenspace of
closed-form linear predictor networks. We show that both avoid collapse through
implicit variance regularization albeit through different dynamical mechanisms.
Moreover, we find that the eigenvalues act as effective learning rate
multipliers and propose a family of isotropic loss functions (IsoLoss) that
equalize convergence rates across eigenmodes. Empirically, IsoLoss speeds up
the initial learning dynamics and increases robustness, thereby allowing us to
dispense with the EMA target network typically used with non-contrastive
methods. Our analysis sheds light on the variance regularization mechanisms of
non-contrastive SSL and lays the theoretical grounds for crafting novel loss
functions that shape the learning dynamics of the predictor's spectrum.</abstract><doi>10.48550/arxiv.2212.04858</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2212.04858 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2212_04858 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Learning Computer Science - Neural and Evolutionary Computing |
title | Implicit variance regularization in non-contrastive SSL |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-06T23%3A53%3A54IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Implicit%20variance%20regularization%20in%20non-contrastive%20SSL&rft.au=Halvagal,%20Manu%20Srinath&rft.date=2022-12-09&rft_id=info:doi/10.48550/arxiv.2212.04858&rft_dat=%3Carxiv_GOX%3E2212_04858%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |