Continuous-Time Meta-Learning with Forward Mode Differentiation

Drawing inspiration from gradient-based meta-learning methods with infinitely small gradient steps, we introduce Continuous-Time Meta-Learning (COMLN), a meta-learning algorithm where adaptation follows the dynamics of a gradient vector field. Specifically, representations of the inputs are meta-lea...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2022-03
Hauptverfasser: Deleu, Tristan, Kanaa, David, Feng, Leo, Kerg, Giancarlo, Bengio, Yoshua, Lajoie, Guillaume, Bacon, Pierre-Luc
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Deleu, Tristan
Kanaa, David
Feng, Leo
Kerg, Giancarlo
Bengio, Yoshua
Lajoie, Guillaume
Bacon, Pierre-Luc
description Drawing inspiration from gradient-based meta-learning methods with infinitely small gradient steps, we introduce Continuous-Time Meta-Learning (COMLN), a meta-learning algorithm where adaptation follows the dynamics of a gradient vector field. Specifically, representations of the inputs are meta-learned such that a task-specific linear classifier is obtained as a solution of an ordinary differential equation (ODE). Treating the learning process as an ODE offers the notable advantage that the length of the trajectory is now continuous, as opposed to a fixed and discrete number of gradient steps. As a consequence, we can optimize the amount of adaptation necessary to solve a new task using stochastic gradient descent, in addition to learning the initial conditions as is standard practice in gradient-based meta-learning. Importantly, in order to compute the exact meta-gradients required for the outer-loop updates, we devise an efficient algorithm based on forward mode differentiation, whose memory requirements do not scale with the length of the learning trajectory, thus allowing longer adaptation in constant memory. We provide analytical guarantees for the stability of COMLN, we show empirically its efficiency in terms of runtime and memory usage, and we illustrate its effectiveness on a range of few-shot image classification problems.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2635744925</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2635744925</sourcerecordid><originalsourceid>FETCH-proquest_journals_26357449253</originalsourceid><addsrcrecordid>eNqNi7sOgjAUQBsTE4nyD02cm-AtBZ0cUOIgGztpwgVLtFf7CL8vgx_gdIZzzoolIOVBHHOADUu9n7Isg6IEpWTCzhXZYGyk6EVrXsgbDFrcUTtr7MhnEx68Jjdr1_OGeuQXMwzocJl0MGR3bD3op8f0xy3b19e2uom3o09EH7qJorOL6qCQqszzEyj5X_UF_uA4yg</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2635744925</pqid></control><display><type>article</type><title>Continuous-Time Meta-Learning with Forward Mode Differentiation</title><source>Ejournal Publishers (free content)</source><creator>Deleu, Tristan ; Kanaa, David ; Feng, Leo ; Kerg, Giancarlo ; Bengio, Yoshua ; Lajoie, Guillaume ; Bacon, Pierre-Luc</creator><creatorcontrib>Deleu, Tristan ; Kanaa, David ; Feng, Leo ; Kerg, Giancarlo ; Bengio, Yoshua ; Lajoie, Guillaume ; Bacon, Pierre-Luc</creatorcontrib><description>Drawing inspiration from gradient-based meta-learning methods with infinitely small gradient steps, we introduce Continuous-Time Meta-Learning (COMLN), a meta-learning algorithm where adaptation follows the dynamics of a gradient vector field. Specifically, representations of the inputs are meta-learned such that a task-specific linear classifier is obtained as a solution of an ordinary differential equation (ODE). Treating the learning process as an ODE offers the notable advantage that the length of the trajectory is now continuous, as opposed to a fixed and discrete number of gradient steps. As a consequence, we can optimize the amount of adaptation necessary to solve a new task using stochastic gradient descent, in addition to learning the initial conditions as is standard practice in gradient-based meta-learning. Importantly, in order to compute the exact meta-gradients required for the outer-loop updates, we devise an efficient algorithm based on forward mode differentiation, whose memory requirements do not scale with the length of the learning trajectory, thus allowing longer adaptation in constant memory. We provide analytical guarantees for the stability of COMLN, we show empirically its efficiency in terms of runtime and memory usage, and we illustrate its effectiveness on a range of few-shot image classification problems.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Adaptation ; Algorithms ; Differential equations ; Differentiation ; Fields (mathematics) ; Image classification ; Initial conditions ; Machine learning ; Ordinary differential equations ; Run time (computers) ; Stability analysis</subject><ispartof>arXiv.org, 2022-03</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>780,784</link.rule.ids></links><search><creatorcontrib>Deleu, Tristan</creatorcontrib><creatorcontrib>Kanaa, David</creatorcontrib><creatorcontrib>Feng, Leo</creatorcontrib><creatorcontrib>Kerg, Giancarlo</creatorcontrib><creatorcontrib>Bengio, Yoshua</creatorcontrib><creatorcontrib>Lajoie, Guillaume</creatorcontrib><creatorcontrib>Bacon, Pierre-Luc</creatorcontrib><title>Continuous-Time Meta-Learning with Forward Mode Differentiation</title><title>arXiv.org</title><description>Drawing inspiration from gradient-based meta-learning methods with infinitely small gradient steps, we introduce Continuous-Time Meta-Learning (COMLN), a meta-learning algorithm where adaptation follows the dynamics of a gradient vector field. Specifically, representations of the inputs are meta-learned such that a task-specific linear classifier is obtained as a solution of an ordinary differential equation (ODE). Treating the learning process as an ODE offers the notable advantage that the length of the trajectory is now continuous, as opposed to a fixed and discrete number of gradient steps. As a consequence, we can optimize the amount of adaptation necessary to solve a new task using stochastic gradient descent, in addition to learning the initial conditions as is standard practice in gradient-based meta-learning. Importantly, in order to compute the exact meta-gradients required for the outer-loop updates, we devise an efficient algorithm based on forward mode differentiation, whose memory requirements do not scale with the length of the learning trajectory, thus allowing longer adaptation in constant memory. We provide analytical guarantees for the stability of COMLN, we show empirically its efficiency in terms of runtime and memory usage, and we illustrate its effectiveness on a range of few-shot image classification problems.</description><subject>Adaptation</subject><subject>Algorithms</subject><subject>Differential equations</subject><subject>Differentiation</subject><subject>Fields (mathematics)</subject><subject>Image classification</subject><subject>Initial conditions</subject><subject>Machine learning</subject><subject>Ordinary differential equations</subject><subject>Run time (computers)</subject><subject>Stability analysis</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNqNi7sOgjAUQBsTE4nyD02cm-AtBZ0cUOIgGztpwgVLtFf7CL8vgx_gdIZzzoolIOVBHHOADUu9n7Isg6IEpWTCzhXZYGyk6EVrXsgbDFrcUTtr7MhnEx68Jjdr1_OGeuQXMwzocJl0MGR3bD3op8f0xy3b19e2uom3o09EH7qJorOL6qCQqszzEyj5X_UF_uA4yg</recordid><startdate>20220302</startdate><enddate>20220302</enddate><creator>Deleu, Tristan</creator><creator>Kanaa, David</creator><creator>Feng, Leo</creator><creator>Kerg, Giancarlo</creator><creator>Bengio, Yoshua</creator><creator>Lajoie, Guillaume</creator><creator>Bacon, Pierre-Luc</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20220302</creationdate><title>Continuous-Time Meta-Learning with Forward Mode Differentiation</title><author>Deleu, Tristan ; Kanaa, David ; Feng, Leo ; Kerg, Giancarlo ; Bengio, Yoshua ; Lajoie, Guillaume ; Bacon, Pierre-Luc</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_26357449253</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Adaptation</topic><topic>Algorithms</topic><topic>Differential equations</topic><topic>Differentiation</topic><topic>Fields (mathematics)</topic><topic>Image classification</topic><topic>Initial conditions</topic><topic>Machine learning</topic><topic>Ordinary differential equations</topic><topic>Run time (computers)</topic><topic>Stability analysis</topic><toplevel>online_resources</toplevel><creatorcontrib>Deleu, Tristan</creatorcontrib><creatorcontrib>Kanaa, David</creatorcontrib><creatorcontrib>Feng, Leo</creatorcontrib><creatorcontrib>Kerg, Giancarlo</creatorcontrib><creatorcontrib>Bengio, Yoshua</creatorcontrib><creatorcontrib>Lajoie, Guillaume</creatorcontrib><creatorcontrib>Bacon, Pierre-Luc</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Access via ProQuest (Open Access)</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Deleu, Tristan</au><au>Kanaa, David</au><au>Feng, Leo</au><au>Kerg, Giancarlo</au><au>Bengio, Yoshua</au><au>Lajoie, Guillaume</au><au>Bacon, Pierre-Luc</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Continuous-Time Meta-Learning with Forward Mode Differentiation</atitle><jtitle>arXiv.org</jtitle><date>2022-03-02</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>Drawing inspiration from gradient-based meta-learning methods with infinitely small gradient steps, we introduce Continuous-Time Meta-Learning (COMLN), a meta-learning algorithm where adaptation follows the dynamics of a gradient vector field. Specifically, representations of the inputs are meta-learned such that a task-specific linear classifier is obtained as a solution of an ordinary differential equation (ODE). Treating the learning process as an ODE offers the notable advantage that the length of the trajectory is now continuous, as opposed to a fixed and discrete number of gradient steps. As a consequence, we can optimize the amount of adaptation necessary to solve a new task using stochastic gradient descent, in addition to learning the initial conditions as is standard practice in gradient-based meta-learning. Importantly, in order to compute the exact meta-gradients required for the outer-loop updates, we devise an efficient algorithm based on forward mode differentiation, whose memory requirements do not scale with the length of the learning trajectory, thus allowing longer adaptation in constant memory. We provide analytical guarantees for the stability of COMLN, we show empirically its efficiency in terms of runtime and memory usage, and we illustrate its effectiveness on a range of few-shot image classification problems.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-03
issn 2331-8422
language eng
recordid cdi_proquest_journals_2635744925
source Ejournal Publishers (free content)
subjects Adaptation
Algorithms
Differential equations
Differentiation
Fields (mathematics)
Image classification
Initial conditions
Machine learning
Ordinary differential equations
Run time (computers)
Stability analysis
title Continuous-Time Meta-Learning with Forward Mode Differentiation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-01T12%3A55%3A19IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Continuous-Time%20Meta-Learning%20with%20Forward%20Mode%20Differentiation&rft.jtitle=arXiv.org&rft.au=Deleu,%20Tristan&rft.date=2022-03-02&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2635744925%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2635744925&rft_id=info:pmid/&rfr_iscdi=true