An algorithmic view of l2 regularization and some path-following algorithms

We establish an equivalence between the l(2)-regularized solution path for a convex loss function, and the solution of an ordinary differentiable equation (ODE). Importantly, this equivalence reveals that the solution path can be viewed as the flow of a hybrid of gradient descent and Newton method a...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of machine learning research 2021-06, Vol.22
Hauptverfasser: Zhu, Yunzhang, Liu, Renxiong
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title Journal of machine learning research
container_volume 22
creator Zhu, Yunzhang
Liu, Renxiong
description We establish an equivalence between the l(2)-regularized solution path for a convex loss function, and the solution of an ordinary differentiable equation (ODE). Importantly, this equivalence reveals that the solution path can be viewed as the flow of a hybrid of gradient descent and Newton method applying to the empirical loss, which is similar to a widely used optimization technique called trust region method. This provides an interesting algorithmic view of l(2) regularization, and is in contrast to the conventional view that the l(2) regularization solution path is similar to the gradient flow of the empirical loss. New path-following algorithms based on homotopy methods and numerical ODE solvers are proposed to numerically approximate the solution path. In particular, we consider respectively Newton method and gradient descent method as the basis algorithm for the homotopy method, and establish their approximation error rates over the solution path. Importantly, our theory suggests novel schemes to choose grid points that guarantee an arbitrarily small suboptimality for the solution path. In terms of computational cost, we prove that in order to achieve an f-suboptimality for the entire solution path, the number of Newton steps required for the Newton method is O(f(-1/2)), while the number of gradient steps required for the gradient descent method is O (f(-1)ln(f(-1))). Finally, we use l(2)-regularized logistic regression as an illustrating example to demonstrate the effectiveness of the proposed path-following algorithms.
format Article
fullrecord <record><control><sourceid>webofscience</sourceid><recordid>TN_cdi_webofscience_primary_000687174600001</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>000687174600001</sourcerecordid><originalsourceid>FETCH-LOGICAL-s155t-96a412c5615d50aad5a247bafcbbf9b710a1a786feb4704597707a525bbc0983</originalsourceid><addsrcrecordid>eNqNjj1PwzAURT2ARCn8B-8oku342clYRXyJSizdq2fHTo2cuIpdIvj1VIDEynTPcM_VvSArDrWopKzhilzn_MYY1yDUirxsJopxSHMohzFY-h7cQpOnUdDZDaeIc_jEEtK5NfU0p9HRI5ZD5VOMaQnT8GfnG3LpMWZ3-5trsnu433VP1fb18bnbbKvMAUrVKpRcWFAcemCIPaCQ2qC3xvjWaM6Qo26Ud0ZqJqHVmmkEAcZY1jb1mjQ_s4szyWcb3GTd_jiHEeePPWNMNZprqc7EeBfK9_0unaZyVu_-r9Zffr5dVw</addsrcrecordid><sourcetype>Enrichment Source</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>An algorithmic view of l2 regularization and some path-following algorithms</title><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><source>ACM Digital Library</source><creator>Zhu, Yunzhang ; Liu, Renxiong</creator><creatorcontrib>Zhu, Yunzhang ; Liu, Renxiong</creatorcontrib><description>We establish an equivalence between the l(2)-regularized solution path for a convex loss function, and the solution of an ordinary differentiable equation (ODE). Importantly, this equivalence reveals that the solution path can be viewed as the flow of a hybrid of gradient descent and Newton method applying to the empirical loss, which is similar to a widely used optimization technique called trust region method. This provides an interesting algorithmic view of l(2) regularization, and is in contrast to the conventional view that the l(2) regularization solution path is similar to the gradient flow of the empirical loss. New path-following algorithms based on homotopy methods and numerical ODE solvers are proposed to numerically approximate the solution path. In particular, we consider respectively Newton method and gradient descent method as the basis algorithm for the homotopy method, and establish their approximation error rates over the solution path. Importantly, our theory suggests novel schemes to choose grid points that guarantee an arbitrarily small suboptimality for the solution path. In terms of computational cost, we prove that in order to achieve an f-suboptimality for the entire solution path, the number of Newton steps required for the Newton method is O(f(-1/2)), while the number of gradient steps required for the gradient descent method is O (f(-1)ln(f(-1))). Finally, we use l(2)-regularized logistic regression as an illustrating example to demonstrate the effectiveness of the proposed path-following algorithms.</description><identifier>ISSN: 1532-4435</identifier><language>eng</language><publisher>BROOKLINE: Microtome Publ</publisher><subject>Automation &amp; Control Systems ; Computer Science ; Computer Science, Artificial Intelligence ; Science &amp; Technology ; Technology</subject><ispartof>Journal of machine learning research, 2021-06, Vol.22</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>true</woscitedreferencessubscribed><woscitedreferencescount>0</woscitedreferencescount><woscitedreferencesoriginalsourcerecordid>wos000687174600001</woscitedreferencesoriginalsourcerecordid><cites>FETCH-LOGICAL-s155t-96a412c5615d50aad5a247bafcbbf9b710a1a786feb4704597707a525bbc0983</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>315,781,785</link.rule.ids></links><search><creatorcontrib>Zhu, Yunzhang</creatorcontrib><creatorcontrib>Liu, Renxiong</creatorcontrib><title>An algorithmic view of l2 regularization and some path-following algorithms</title><title>Journal of machine learning research</title><addtitle>J MACH LEARN RES</addtitle><description>We establish an equivalence between the l(2)-regularized solution path for a convex loss function, and the solution of an ordinary differentiable equation (ODE). Importantly, this equivalence reveals that the solution path can be viewed as the flow of a hybrid of gradient descent and Newton method applying to the empirical loss, which is similar to a widely used optimization technique called trust region method. This provides an interesting algorithmic view of l(2) regularization, and is in contrast to the conventional view that the l(2) regularization solution path is similar to the gradient flow of the empirical loss. New path-following algorithms based on homotopy methods and numerical ODE solvers are proposed to numerically approximate the solution path. In particular, we consider respectively Newton method and gradient descent method as the basis algorithm for the homotopy method, and establish their approximation error rates over the solution path. Importantly, our theory suggests novel schemes to choose grid points that guarantee an arbitrarily small suboptimality for the solution path. In terms of computational cost, we prove that in order to achieve an f-suboptimality for the entire solution path, the number of Newton steps required for the Newton method is O(f(-1/2)), while the number of gradient steps required for the gradient descent method is O (f(-1)ln(f(-1))). Finally, we use l(2)-regularized logistic regression as an illustrating example to demonstrate the effectiveness of the proposed path-following algorithms.</description><subject>Automation &amp; Control Systems</subject><subject>Computer Science</subject><subject>Computer Science, Artificial Intelligence</subject><subject>Science &amp; Technology</subject><subject>Technology</subject><issn>1532-4435</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>HGBXW</sourceid><recordid>eNqNjj1PwzAURT2ARCn8B-8oku342clYRXyJSizdq2fHTo2cuIpdIvj1VIDEynTPcM_VvSArDrWopKzhilzn_MYY1yDUirxsJopxSHMohzFY-h7cQpOnUdDZDaeIc_jEEtK5NfU0p9HRI5ZD5VOMaQnT8GfnG3LpMWZ3-5trsnu433VP1fb18bnbbKvMAUrVKpRcWFAcemCIPaCQ2qC3xvjWaM6Qo26Ud0ZqJqHVmmkEAcZY1jb1mjQ_s4szyWcb3GTd_jiHEeePPWNMNZprqc7EeBfK9_0unaZyVu_-r9Zffr5dVw</recordid><startdate>20210601</startdate><enddate>20210601</enddate><creator>Zhu, Yunzhang</creator><creator>Liu, Renxiong</creator><general>Microtome Publ</general><scope>BLEPL</scope><scope>DTL</scope><scope>HGBXW</scope></search><sort><creationdate>20210601</creationdate><title>An algorithmic view of l2 regularization and some path-following algorithms</title><author>Zhu, Yunzhang ; Liu, Renxiong</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-s155t-96a412c5615d50aad5a247bafcbbf9b710a1a786feb4704597707a525bbc0983</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Automation &amp; Control Systems</topic><topic>Computer Science</topic><topic>Computer Science, Artificial Intelligence</topic><topic>Science &amp; Technology</topic><topic>Technology</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhu, Yunzhang</creatorcontrib><creatorcontrib>Liu, Renxiong</creatorcontrib><collection>Web of Science Core Collection</collection><collection>Science Citation Index Expanded</collection><collection>Web of Science - Science Citation Index Expanded - 2021</collection><jtitle>Journal of machine learning research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhu, Yunzhang</au><au>Liu, Renxiong</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>An algorithmic view of l2 regularization and some path-following algorithms</atitle><jtitle>Journal of machine learning research</jtitle><stitle>J MACH LEARN RES</stitle><date>2021-06-01</date><risdate>2021</risdate><volume>22</volume><issn>1532-4435</issn><abstract>We establish an equivalence between the l(2)-regularized solution path for a convex loss function, and the solution of an ordinary differentiable equation (ODE). Importantly, this equivalence reveals that the solution path can be viewed as the flow of a hybrid of gradient descent and Newton method applying to the empirical loss, which is similar to a widely used optimization technique called trust region method. This provides an interesting algorithmic view of l(2) regularization, and is in contrast to the conventional view that the l(2) regularization solution path is similar to the gradient flow of the empirical loss. New path-following algorithms based on homotopy methods and numerical ODE solvers are proposed to numerically approximate the solution path. In particular, we consider respectively Newton method and gradient descent method as the basis algorithm for the homotopy method, and establish their approximation error rates over the solution path. Importantly, our theory suggests novel schemes to choose grid points that guarantee an arbitrarily small suboptimality for the solution path. In terms of computational cost, we prove that in order to achieve an f-suboptimality for the entire solution path, the number of Newton steps required for the Newton method is O(f(-1/2)), while the number of gradient steps required for the gradient descent method is O (f(-1)ln(f(-1))). Finally, we use l(2)-regularized logistic regression as an illustrating example to demonstrate the effectiveness of the proposed path-following algorithms.</abstract><cop>BROOKLINE</cop><pub>Microtome Publ</pub><tpages>62</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1532-4435
ispartof Journal of machine learning research, 2021-06, Vol.22
issn 1532-4435
language eng
recordid cdi_webofscience_primary_000687174600001
source Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals; ACM Digital Library
subjects Automation & Control Systems
Computer Science
Computer Science, Artificial Intelligence
Science & Technology
Technology
title An algorithmic view of l2 regularization and some path-following algorithms
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-17T03%3A53%3A00IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-webofscience&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=An%20algorithmic%20view%20of%20l2%20regularization%20and%20some%20path-following%20algorithms&rft.jtitle=Journal%20of%20machine%20learning%20research&rft.au=Zhu,%20Yunzhang&rft.date=2021-06-01&rft.volume=22&rft.issn=1532-4435&rft_id=info:doi/&rft_dat=%3Cwebofscience%3E000687174600001%3C/webofscience%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true