Expressivity of ReLU-Networks under Convex Relaxations

Convex relaxations are a key component of training and certifying provably safe neural networks. However, despite substantial progress, a wide and poorly understood accuracy gap to standard networks remains, raising the question of whether this is due to fundamental limitations of convex relaxations...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Baader, Maximilian, Müller, Mark Niklas, Mao, Yuhao, Vechev, Martin
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Baader, Maximilian
Müller, Mark Niklas
Mao, Yuhao
Vechev, Martin
description Convex relaxations are a key component of training and certifying provably safe neural networks. However, despite substantial progress, a wide and poorly understood accuracy gap to standard networks remains, raising the question of whether this is due to fundamental limitations of convex relaxations. Initial work investigating this question focused on the simple and widely used IBP relaxation. It revealed that some univariate, convex, continuous piecewise linear (CPWL) functions cannot be encoded by any ReLU network such that its IBP-analysis is precise. To explore whether this limitation is shared by more advanced convex relaxations, we conduct the first in-depth study on the expressive power of ReLU networks across all commonly used convex relaxations. We show that: (i) more advanced relaxations allow a larger class of univariate functions to be expressed as precisely analyzable ReLU networks, (ii) more precise relaxations can allow exponentially larger solution spaces of ReLU networks encoding the same functions, and (iii) even using the most precise single-neuron relaxations, it is impossible to construct precisely analyzable ReLU networks that express multivariate, convex, monotone CPWL functions.
doi_str_mv 10.48550/arxiv.2311.04015
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_04015</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_04015</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-8ecae2450d3a6e96a48faaf0321d9344dec1e5e22a8e6094ec36827b7965a0b33</originalsourceid><addsrcrecordid>eNotj81qg0AURmfTRbF9gK7qC2jm33EZxLYBaSCYtVz1CkONyoyx-vZN067O4oPDdwh5YTSWRim6A7faJeaCsZhKytQj0fk6OfTeLnbewrELT1ico0-cv0f35cPr0KILs3FYcL1NPaww23HwT-Shg97j8z8DUr7lZfYRFcf3Q7YvItCJigw2gFwq2grQmGqQpgPoqOCsTYWULTYMFXIOBjVNJTZCG57USaoV0FqIgLz-ae_Hq8nZC7it-g2o7gHiBzSyQGE</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Expressivity of ReLU-Networks under Convex Relaxations</title><source>arXiv.org</source><creator>Baader, Maximilian ; Müller, Mark Niklas ; Mao, Yuhao ; Vechev, Martin</creator><creatorcontrib>Baader, Maximilian ; Müller, Mark Niklas ; Mao, Yuhao ; Vechev, Martin</creatorcontrib><description>Convex relaxations are a key component of training and certifying provably safe neural networks. However, despite substantial progress, a wide and poorly understood accuracy gap to standard networks remains, raising the question of whether this is due to fundamental limitations of convex relaxations. Initial work investigating this question focused on the simple and widely used IBP relaxation. It revealed that some univariate, convex, continuous piecewise linear (CPWL) functions cannot be encoded by any ReLU network such that its IBP-analysis is precise. To explore whether this limitation is shared by more advanced convex relaxations, we conduct the first in-depth study on the expressive power of ReLU networks across all commonly used convex relaxations. We show that: (i) more advanced relaxations allow a larger class of univariate functions to be expressed as precisely analyzable ReLU networks, (ii) more precise relaxations can allow exponentially larger solution spaces of ReLU networks encoding the same functions, and (iii) even using the most precise single-neuron relaxations, it is impossible to construct precisely analyzable ReLU networks that express multivariate, convex, monotone CPWL functions.</description><identifier>DOI: 10.48550/arxiv.2311.04015</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Learning</subject><creationdate>2023-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.04015$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.04015$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Baader, Maximilian</creatorcontrib><creatorcontrib>Müller, Mark Niklas</creatorcontrib><creatorcontrib>Mao, Yuhao</creatorcontrib><creatorcontrib>Vechev, Martin</creatorcontrib><title>Expressivity of ReLU-Networks under Convex Relaxations</title><description>Convex relaxations are a key component of training and certifying provably safe neural networks. However, despite substantial progress, a wide and poorly understood accuracy gap to standard networks remains, raising the question of whether this is due to fundamental limitations of convex relaxations. Initial work investigating this question focused on the simple and widely used IBP relaxation. It revealed that some univariate, convex, continuous piecewise linear (CPWL) functions cannot be encoded by any ReLU network such that its IBP-analysis is precise. To explore whether this limitation is shared by more advanced convex relaxations, we conduct the first in-depth study on the expressive power of ReLU networks across all commonly used convex relaxations. We show that: (i) more advanced relaxations allow a larger class of univariate functions to be expressed as precisely analyzable ReLU networks, (ii) more precise relaxations can allow exponentially larger solution spaces of ReLU networks encoding the same functions, and (iii) even using the most precise single-neuron relaxations, it is impossible to construct precisely analyzable ReLU networks that express multivariate, convex, monotone CPWL functions.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81qg0AURmfTRbF9gK7qC2jm33EZxLYBaSCYtVz1CkONyoyx-vZN067O4oPDdwh5YTSWRim6A7faJeaCsZhKytQj0fk6OfTeLnbewrELT1ico0-cv0f35cPr0KILs3FYcL1NPaww23HwT-Shg97j8z8DUr7lZfYRFcf3Q7YvItCJigw2gFwq2grQmGqQpgPoqOCsTYWULTYMFXIOBjVNJTZCG57USaoV0FqIgLz-ae_Hq8nZC7it-g2o7gHiBzSyQGE</recordid><startdate>20231107</startdate><enddate>20231107</enddate><creator>Baader, Maximilian</creator><creator>Müller, Mark Niklas</creator><creator>Mao, Yuhao</creator><creator>Vechev, Martin</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231107</creationdate><title>Expressivity of ReLU-Networks under Convex Relaxations</title><author>Baader, Maximilian ; Müller, Mark Niklas ; Mao, Yuhao ; Vechev, Martin</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-8ecae2450d3a6e96a48faaf0321d9344dec1e5e22a8e6094ec36827b7965a0b33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Baader, Maximilian</creatorcontrib><creatorcontrib>Müller, Mark Niklas</creatorcontrib><creatorcontrib>Mao, Yuhao</creatorcontrib><creatorcontrib>Vechev, Martin</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Baader, Maximilian</au><au>Müller, Mark Niklas</au><au>Mao, Yuhao</au><au>Vechev, Martin</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Expressivity of ReLU-Networks under Convex Relaxations</atitle><date>2023-11-07</date><risdate>2023</risdate><abstract>Convex relaxations are a key component of training and certifying provably safe neural networks. However, despite substantial progress, a wide and poorly understood accuracy gap to standard networks remains, raising the question of whether this is due to fundamental limitations of convex relaxations. Initial work investigating this question focused on the simple and widely used IBP relaxation. It revealed that some univariate, convex, continuous piecewise linear (CPWL) functions cannot be encoded by any ReLU network such that its IBP-analysis is precise. To explore whether this limitation is shared by more advanced convex relaxations, we conduct the first in-depth study on the expressive power of ReLU networks across all commonly used convex relaxations. We show that: (i) more advanced relaxations allow a larger class of univariate functions to be expressed as precisely analyzable ReLU networks, (ii) more precise relaxations can allow exponentially larger solution spaces of ReLU networks encoding the same functions, and (iii) even using the most precise single-neuron relaxations, it is impossible to construct precisely analyzable ReLU networks that express multivariate, convex, monotone CPWL functions.</abstract><doi>10.48550/arxiv.2311.04015</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2311.04015
ispartof
issn
language eng
recordid cdi_arxiv_primary_2311_04015
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Learning
title Expressivity of ReLU-Networks under Convex Relaxations
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-11T11%3A00%3A57IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Expressivity%20of%20ReLU-Networks%20under%20Convex%20Relaxations&rft.au=Baader,%20Maximilian&rft.date=2023-11-07&rft_id=info:doi/10.48550/arxiv.2311.04015&rft_dat=%3Carxiv_GOX%3E2311_04015%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true