Distributed Continual Learning with CoCoA in High-dimensional Linear Regression

We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Hellkvist, Martin, Özçelikkale, Ayça, Ahlén, Anders
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Hellkvist, Martin
Özçelikkale, Ayça
Ahlén, Anders
description We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly arrived task without performance degradation on the previously seen tasks. In contrast to the continual learning literature focusing on the centralized setting, we investigate the problem from a distributed estimation perspective. We consider the well-established distributed learning algorithm COCOA, which distributes the model parameters and the corresponding features over the network. We provide exact analytical characterization for the generalization error of COCOA under continual learning for linear regression in a range of scenarios, where overparameterization is of particular interest. These analytical results characterize how the generalization error depends on the network structure, the task similarity and the number of tasks, and show how these dependencies are intertwined. In particular, our results show that the generalization error can be significantly reduced by adjusting the network size, where the most favorable network size depends on task similarity and the number of tasks. We present numerical results verifying the theoretical analysis and illustrate the continual learning performance of COCOA with a digit classification task.
doi_str_mv 10.48550/arxiv.2312.01795
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2312_01795</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2312_01795</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-28ebb54369a016272531025d49e17bc48972fbe4323bd29d6ec43b19e3b950673</originalsourceid><addsrcrecordid>eNotj81qwkAURmfTRbF9gK46L5B0_iezlFRrISCI-zDjXOMFnZRJ1Pbta2xXHxwOHxxCXjgrVaU1e_P5Gy-lkFyUjFunH8n6HYcxYziPEGndpxHT2R9pAz4nTB294ni48bqfU0x0hd2hiHiCNGCfJg_TzaQb6DIME3siD3t_HOD5f2dku1xs61XRrD8-63lTeGN1ISoIQStpnGfcCCu05EzoqBxwG3aqclbsAygpZIjCRQM7JQN3IIPTzFg5I69_t_ei9ivjyeefdipr72XyF6S9SDs</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Distributed Continual Learning with CoCoA in High-dimensional Linear Regression</title><source>arXiv.org</source><creator>Hellkvist, Martin ; Özçelikkale, Ayça ; Ahlén, Anders</creator><creatorcontrib>Hellkvist, Martin ; Özçelikkale, Ayça ; Ahlén, Anders</creatorcontrib><description>We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly arrived task without performance degradation on the previously seen tasks. In contrast to the continual learning literature focusing on the centralized setting, we investigate the problem from a distributed estimation perspective. We consider the well-established distributed learning algorithm COCOA, which distributes the model parameters and the corresponding features over the network. We provide exact analytical characterization for the generalization error of COCOA under continual learning for linear regression in a range of scenarios, where overparameterization is of particular interest. These analytical results characterize how the generalization error depends on the network structure, the task similarity and the number of tasks, and show how these dependencies are intertwined. In particular, our results show that the generalization error can be significantly reduced by adjusting the network size, where the most favorable network size depends on task similarity and the number of tasks. We present numerical results verifying the theoretical analysis and illustrate the continual learning performance of COCOA with a digit classification task.</description><identifier>DOI: 10.48550/arxiv.2312.01795</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2023-12</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2312.01795$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2312.01795$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Hellkvist, Martin</creatorcontrib><creatorcontrib>Özçelikkale, Ayça</creatorcontrib><creatorcontrib>Ahlén, Anders</creatorcontrib><title>Distributed Continual Learning with CoCoA in High-dimensional Linear Regression</title><description>We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly arrived task without performance degradation on the previously seen tasks. In contrast to the continual learning literature focusing on the centralized setting, we investigate the problem from a distributed estimation perspective. We consider the well-established distributed learning algorithm COCOA, which distributes the model parameters and the corresponding features over the network. We provide exact analytical characterization for the generalization error of COCOA under continual learning for linear regression in a range of scenarios, where overparameterization is of particular interest. These analytical results characterize how the generalization error depends on the network structure, the task similarity and the number of tasks, and show how these dependencies are intertwined. In particular, our results show that the generalization error can be significantly reduced by adjusting the network size, where the most favorable network size depends on task similarity and the number of tasks. We present numerical results verifying the theoretical analysis and illustrate the continual learning performance of COCOA with a digit classification task.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81qwkAURmfTRbF9gK46L5B0_iezlFRrISCI-zDjXOMFnZRJ1Pbta2xXHxwOHxxCXjgrVaU1e_P5Gy-lkFyUjFunH8n6HYcxYziPEGndpxHT2R9pAz4nTB294ni48bqfU0x0hd2hiHiCNGCfJg_TzaQb6DIME3siD3t_HOD5f2dku1xs61XRrD8-63lTeGN1ISoIQStpnGfcCCu05EzoqBxwG3aqclbsAygpZIjCRQM7JQN3IIPTzFg5I69_t_ei9ivjyeefdipr72XyF6S9SDs</recordid><startdate>20231204</startdate><enddate>20231204</enddate><creator>Hellkvist, Martin</creator><creator>Özçelikkale, Ayça</creator><creator>Ahlén, Anders</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231204</creationdate><title>Distributed Continual Learning with CoCoA in High-dimensional Linear Regression</title><author>Hellkvist, Martin ; Özçelikkale, Ayça ; Ahlén, Anders</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-28ebb54369a016272531025d49e17bc48972fbe4323bd29d6ec43b19e3b950673</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Hellkvist, Martin</creatorcontrib><creatorcontrib>Özçelikkale, Ayça</creatorcontrib><creatorcontrib>Ahlén, Anders</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Hellkvist, Martin</au><au>Özçelikkale, Ayça</au><au>Ahlén, Anders</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Distributed Continual Learning with CoCoA in High-dimensional Linear Regression</atitle><date>2023-12-04</date><risdate>2023</risdate><abstract>We consider estimation under scenarios where the signals of interest exhibit change of characteristics over time. In particular, we consider the continual learning problem where different tasks, e.g., data with different distributions, arrive sequentially and the aim is to perform well on the newly arrived task without performance degradation on the previously seen tasks. In contrast to the continual learning literature focusing on the centralized setting, we investigate the problem from a distributed estimation perspective. We consider the well-established distributed learning algorithm COCOA, which distributes the model parameters and the corresponding features over the network. We provide exact analytical characterization for the generalization error of COCOA under continual learning for linear regression in a range of scenarios, where overparameterization is of particular interest. These analytical results characterize how the generalization error depends on the network structure, the task similarity and the number of tasks, and show how these dependencies are intertwined. In particular, our results show that the generalization error can be significantly reduced by adjusting the network size, where the most favorable network size depends on task similarity and the number of tasks. We present numerical results verifying the theoretical analysis and illustrate the continual learning performance of COCOA with a digit classification task.</abstract><doi>10.48550/arxiv.2312.01795</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2312.01795
ispartof
issn
language eng
recordid cdi_arxiv_primary_2312_01795
source arXiv.org
subjects Computer Science - Learning
title Distributed Continual Learning with CoCoA in High-dimensional Linear Regression
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T15%3A24%3A32IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Distributed%20Continual%20Learning%20with%20CoCoA%20in%20High-dimensional%20Linear%20Regression&rft.au=Hellkvist,%20Martin&rft.date=2023-12-04&rft_id=info:doi/10.48550/arxiv.2312.01795&rft_dat=%3Carxiv_GOX%3E2312_01795%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true