A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming

Bilevel programming has recently received attention in the literature due to its wide range of applications, including reinforcement learning and hyper-parameter optimization. However, it is widely assumed that the underlying bilevel optimization problem is solved either by a single machine or, in t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-10
Hauptverfasser: Nazari, Parvin, Mousavi, Ahmad, Davoud Ataee Tarzanagh, Michailidis, George
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Nazari, Parvin
Mousavi, Ahmad
Davoud Ataee Tarzanagh
Michailidis, George
description Bilevel programming has recently received attention in the literature due to its wide range of applications, including reinforcement learning and hyper-parameter optimization. However, it is widely assumed that the underlying bilevel optimization problem is solved either by a single machine or, in the case of multiple machines connected in a star-shaped network, i.e., in a federated learning setting. The latter approach suffers from a high communication cost on the central node (e.g., parameter server). Hence, there is an interest in developing methods that solve bilevel optimization problems in a communication-efficient, decentralized manner. To that end, this paper introduces a penalty function-based decentralized algorithm with theoretical guarantees for this class of optimization problems. Specifically, a distributed alternating gradient-type algorithm for solving consensus bilevel programming over a decentralized network is developed. A key feature of the proposed algorithm is the estimation of the hyper-gradient of the penalty function through decentralized computation of matrix-vector products and a few vector communications. The estimation is integrated into an alternating algorithm for solving the penalized reformulation of the bilevel optimization problem. Under appropriate step sizes and penalty parameters, our theoretical framework ensures non-asymptotic convergence to the optimal solution of the original problem under various convexity conditions. Our theoretical result highlights improvements in the iteration complexity of decentralized bilevel optimization, all while making efficient use of vector communication. Empirical results demonstrate that the proposed method performs well in real-world settings.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2734429176</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2734429176</sourcerecordid><originalsourceid>FETCH-proquest_journals_27344291763</originalsourceid><addsrcrecordid>eNqNjb0KwjAYAIMgWLTvEHAutEl_dLS14iJ06F5C-6Wm5EeTVNCnt4MP4HTLHbdCAaE0iQ4pIRsUOjfFcUzygmQZDVB7wg1oJv07KpmDAd_A382AubG4MkrNWvTMC6OjmnPRC9Aen6FfYJkUnyUohYQXSNxYM1qmlNDjDq05kw7CH7dof6nb6ho9rHnO4Hw3mdkuV9eRgqYpOSZFTv-zvl4vQHk</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2734429176</pqid></control><display><type>article</type><title>A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming</title><source>Free E- Journals</source><creator>Nazari, Parvin ; Mousavi, Ahmad ; Davoud Ataee Tarzanagh ; Michailidis, George</creator><creatorcontrib>Nazari, Parvin ; Mousavi, Ahmad ; Davoud Ataee Tarzanagh ; Michailidis, George</creatorcontrib><description>Bilevel programming has recently received attention in the literature due to its wide range of applications, including reinforcement learning and hyper-parameter optimization. However, it is widely assumed that the underlying bilevel optimization problem is solved either by a single machine or, in the case of multiple machines connected in a star-shaped network, i.e., in a federated learning setting. The latter approach suffers from a high communication cost on the central node (e.g., parameter server). Hence, there is an interest in developing methods that solve bilevel optimization problems in a communication-efficient, decentralized manner. To that end, this paper introduces a penalty function-based decentralized algorithm with theoretical guarantees for this class of optimization problems. Specifically, a distributed alternating gradient-type algorithm for solving consensus bilevel programming over a decentralized network is developed. A key feature of the proposed algorithm is the estimation of the hyper-gradient of the penalty function through decentralized computation of matrix-vector products and a few vector communications. The estimation is integrated into an alternating algorithm for solving the penalized reformulation of the bilevel optimization problem. Under appropriate step sizes and penalty parameters, our theoretical framework ensures non-asymptotic convergence to the optimal solution of the original problem under various convexity conditions. Our theoretical result highlights improvements in the iteration complexity of decentralized bilevel optimization, all while making efficient use of vector communication. Empirical results demonstrate that the proposed method performs well in real-world settings.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Algorithms ; Communication ; Convergence ; Convexity ; Empirical analysis ; Machine learning ; Mathematical analysis ; Minimax technique ; Optimization ; Parameters ; Penalty function ; Programming</subject><ispartof>arXiv.org, 2024-10</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by-nc-sa/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Nazari, Parvin</creatorcontrib><creatorcontrib>Mousavi, Ahmad</creatorcontrib><creatorcontrib>Davoud Ataee Tarzanagh</creatorcontrib><creatorcontrib>Michailidis, George</creatorcontrib><title>A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming</title><title>arXiv.org</title><description>Bilevel programming has recently received attention in the literature due to its wide range of applications, including reinforcement learning and hyper-parameter optimization. However, it is widely assumed that the underlying bilevel optimization problem is solved either by a single machine or, in the case of multiple machines connected in a star-shaped network, i.e., in a federated learning setting. The latter approach suffers from a high communication cost on the central node (e.g., parameter server). Hence, there is an interest in developing methods that solve bilevel optimization problems in a communication-efficient, decentralized manner. To that end, this paper introduces a penalty function-based decentralized algorithm with theoretical guarantees for this class of optimization problems. Specifically, a distributed alternating gradient-type algorithm for solving consensus bilevel programming over a decentralized network is developed. A key feature of the proposed algorithm is the estimation of the hyper-gradient of the penalty function through decentralized computation of matrix-vector products and a few vector communications. The estimation is integrated into an alternating algorithm for solving the penalized reformulation of the bilevel optimization problem. Under appropriate step sizes and penalty parameters, our theoretical framework ensures non-asymptotic convergence to the optimal solution of the original problem under various convexity conditions. Our theoretical result highlights improvements in the iteration complexity of decentralized bilevel optimization, all while making efficient use of vector communication. Empirical results demonstrate that the proposed method performs well in real-world settings.</description><subject>Algorithms</subject><subject>Communication</subject><subject>Convergence</subject><subject>Convexity</subject><subject>Empirical analysis</subject><subject>Machine learning</subject><subject>Mathematical analysis</subject><subject>Minimax technique</subject><subject>Optimization</subject><subject>Parameters</subject><subject>Penalty function</subject><subject>Programming</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNqNjb0KwjAYAIMgWLTvEHAutEl_dLS14iJ06F5C-6Wm5EeTVNCnt4MP4HTLHbdCAaE0iQ4pIRsUOjfFcUzygmQZDVB7wg1oJv07KpmDAd_A382AubG4MkrNWvTMC6OjmnPRC9Aen6FfYJkUnyUohYQXSNxYM1qmlNDjDq05kw7CH7dof6nb6ho9rHnO4Hw3mdkuV9eRgqYpOSZFTv-zvl4vQHk</recordid><startdate>20241010</startdate><enddate>20241010</enddate><creator>Nazari, Parvin</creator><creator>Mousavi, Ahmad</creator><creator>Davoud Ataee Tarzanagh</creator><creator>Michailidis, George</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241010</creationdate><title>A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming</title><author>Nazari, Parvin ; Mousavi, Ahmad ; Davoud Ataee Tarzanagh ; Michailidis, George</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27344291763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Communication</topic><topic>Convergence</topic><topic>Convexity</topic><topic>Empirical analysis</topic><topic>Machine learning</topic><topic>Mathematical analysis</topic><topic>Minimax technique</topic><topic>Optimization</topic><topic>Parameters</topic><topic>Penalty function</topic><topic>Programming</topic><toplevel>online_resources</toplevel><creatorcontrib>Nazari, Parvin</creatorcontrib><creatorcontrib>Mousavi, Ahmad</creatorcontrib><creatorcontrib>Davoud Ataee Tarzanagh</creatorcontrib><creatorcontrib>Michailidis, George</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nazari, Parvin</au><au>Mousavi, Ahmad</au><au>Davoud Ataee Tarzanagh</au><au>Michailidis, George</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming</atitle><jtitle>arXiv.org</jtitle><date>2024-10-10</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Bilevel programming has recently received attention in the literature due to its wide range of applications, including reinforcement learning and hyper-parameter optimization. However, it is widely assumed that the underlying bilevel optimization problem is solved either by a single machine or, in the case of multiple machines connected in a star-shaped network, i.e., in a federated learning setting. The latter approach suffers from a high communication cost on the central node (e.g., parameter server). Hence, there is an interest in developing methods that solve bilevel optimization problems in a communication-efficient, decentralized manner. To that end, this paper introduces a penalty function-based decentralized algorithm with theoretical guarantees for this class of optimization problems. Specifically, a distributed alternating gradient-type algorithm for solving consensus bilevel programming over a decentralized network is developed. A key feature of the proposed algorithm is the estimation of the hyper-gradient of the penalty function through decentralized computation of matrix-vector products and a few vector communications. The estimation is integrated into an alternating algorithm for solving the penalized reformulation of the bilevel optimization problem. Under appropriate step sizes and penalty parameters, our theoretical framework ensures non-asymptotic convergence to the optimal solution of the original problem under various convexity conditions. Our theoretical result highlights improvements in the iteration complexity of decentralized bilevel optimization, all while making efficient use of vector communication. Empirical results demonstrate that the proposed method performs well in real-world settings.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-10
issn 2331-8422
language eng
recordid cdi_proquest_journals_2734429176
source Free E- Journals
subjects Algorithms
Communication
Convergence
Convexity
Empirical analysis
Machine learning
Mathematical analysis
Minimax technique
Optimization
Parameters
Penalty function
Programming
title A Penalty-Based Method for Communication-Efficient Decentralized Bilevel Programming
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T14%3A23%3A22IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=A%20Penalty-Based%20Method%20for%20Communication-Efficient%20Decentralized%20Bilevel%20Programming&rft.jtitle=arXiv.org&rft.au=Nazari,%20Parvin&rft.date=2024-10-10&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2734429176%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2734429176&rft_id=info:pmid/&rfr_iscdi=true