Regularizers for structured sparsity

We study the problem of learning a sparse linear regression vector under additional conditions on the structure of its sparsity pattern. This problem is relevant in machine learning, statistics and signal processing. It is well known that a linear regression can benefit from knowledge that the under...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Advances in computational mathematics 2013-04, Vol.38 (3), p.455-489
Hauptverfasser: Micchelli, Charles A., Morales, Jean M., Pontil, Massimiliano
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 489
container_issue 3
container_start_page 455
container_title Advances in computational mathematics
container_volume 38
creator Micchelli, Charles A.
Morales, Jean M.
Pontil, Massimiliano
description We study the problem of learning a sparse linear regression vector under additional conditions on the structure of its sparsity pattern. This problem is relevant in machine learning, statistics and signal processing. It is well known that a linear regression can benefit from knowledge that the underlying regression vector is sparse. The combinatorial problem of selecting the nonzero components of this vector can be “relaxed” by regularizing the squared error with a convex penalty function like the ℓ 1 norm. However, in many applications, additional conditions on the structure of the regression vector and its sparsity pattern are available. Incorporating this information into the learning method may lead to a significant decrease of the estimation error. In this paper, we present a family of convex penalty functions, which encode prior knowledge on the structure of the vector formed by the absolute values of the regression coefficients. This family subsumes the ℓ 1 norm and is flexible enough to include different models of sparsity patterns, which are of practical and theoretical importance. We establish the basic properties of these penalty functions and discuss some examples where they can be computed explicitly. Moreover, we present a convergent optimization algorithm for solving regularized least squares with these penalty functions. Numerical simulations highlight the benefit of structured sparsity and the advantage offered by our approach over the Lasso method and other related methods.
doi_str_mv 10.1007/s10444-011-9245-9
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_miscellaneous_1429908577</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1429908577</sourcerecordid><originalsourceid>FETCH-LOGICAL-c321t-fff4cc249d8cc91153f97a490e1b59415e754cb69c83c46c007bf3f8a039b9f23</originalsourceid><addsrcrecordid>eNp9kMFOwzAMhiMEEmPwANx64MAlEKdJUx_RBANpEhKCc5RmydSpW0fcHsbTk6mcOdmy_s-yP8ZuQTyAEOaRQCiluADgKJXmeMZmoI3kmOfnuReA3EBVX7Iroq0QAiujZ-zuI2zGzqX2JyQqYp8KGtLohzGFdUEHl6gdjtfsIrqOws1fnbOvl-fPxStfvS_fFk8r7ksJA48xKu-lwnXtPQLoMqJxCkWARqMCHYxWvqnQ16VXlc93N7GMtRMlNhhlOWf3095D6r_HQIPdteRD17l96EeyoCSiqLUxOQpT1KeeKIVoD6nduXS0IOzJiJ2M2GzEnoxYzIycGMrZ_SYku-3HtM8f_QP9AksfYtU</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>1429908577</pqid></control><display><type>article</type><title>Regularizers for structured sparsity</title><source>SpringerLink Journals - AutoHoldings</source><creator>Micchelli, Charles A. ; Morales, Jean M. ; Pontil, Massimiliano</creator><creatorcontrib>Micchelli, Charles A. ; Morales, Jean M. ; Pontil, Massimiliano</creatorcontrib><description>We study the problem of learning a sparse linear regression vector under additional conditions on the structure of its sparsity pattern. This problem is relevant in machine learning, statistics and signal processing. It is well known that a linear regression can benefit from knowledge that the underlying regression vector is sparse. The combinatorial problem of selecting the nonzero components of this vector can be “relaxed” by regularizing the squared error with a convex penalty function like the ℓ 1 norm. However, in many applications, additional conditions on the structure of the regression vector and its sparsity pattern are available. Incorporating this information into the learning method may lead to a significant decrease of the estimation error. In this paper, we present a family of convex penalty functions, which encode prior knowledge on the structure of the vector formed by the absolute values of the regression coefficients. This family subsumes the ℓ 1 norm and is flexible enough to include different models of sparsity patterns, which are of practical and theoretical importance. We establish the basic properties of these penalty functions and discuss some examples where they can be computed explicitly. Moreover, we present a convergent optimization algorithm for solving regularized least squares with these penalty functions. Numerical simulations highlight the benefit of structured sparsity and the advantage offered by our approach over the Lasso method and other related methods.</description><identifier>ISSN: 1019-7168</identifier><identifier>EISSN: 1572-9044</identifier><identifier>DOI: 10.1007/s10444-011-9245-9</identifier><language>eng</language><publisher>Boston: Springer US</publisher><subject>Combinatorial analysis ; Computational Mathematics and Numerical Analysis ; Computational Science and Engineering ; Learning ; Mathematical analysis ; Mathematical and Computational Biology ; Mathematical Modeling and Industrial Mathematics ; Mathematical models ; Mathematics ; Mathematics and Statistics ; Norms ; Penalty function ; Regression ; Vectors (mathematics) ; Visualization</subject><ispartof>Advances in computational mathematics, 2013-04, Vol.38 (3), p.455-489</ispartof><rights>Springer Science+Business Media, LLC. 2011</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c321t-fff4cc249d8cc91153f97a490e1b59415e754cb69c83c46c007bf3f8a039b9f23</citedby><cites>FETCH-LOGICAL-c321t-fff4cc249d8cc91153f97a490e1b59415e754cb69c83c46c007bf3f8a039b9f23</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktopdf>$$Uhttps://link.springer.com/content/pdf/10.1007/s10444-011-9245-9$$EPDF$$P50$$Gspringer$$H</linktopdf><linktohtml>$$Uhttps://link.springer.com/10.1007/s10444-011-9245-9$$EHTML$$P50$$Gspringer$$H</linktohtml><link.rule.ids>314,780,784,27924,27925,41488,42557,51319</link.rule.ids></links><search><creatorcontrib>Micchelli, Charles A.</creatorcontrib><creatorcontrib>Morales, Jean M.</creatorcontrib><creatorcontrib>Pontil, Massimiliano</creatorcontrib><title>Regularizers for structured sparsity</title><title>Advances in computational mathematics</title><addtitle>Adv Comput Math</addtitle><description>We study the problem of learning a sparse linear regression vector under additional conditions on the structure of its sparsity pattern. This problem is relevant in machine learning, statistics and signal processing. It is well known that a linear regression can benefit from knowledge that the underlying regression vector is sparse. The combinatorial problem of selecting the nonzero components of this vector can be “relaxed” by regularizing the squared error with a convex penalty function like the ℓ 1 norm. However, in many applications, additional conditions on the structure of the regression vector and its sparsity pattern are available. Incorporating this information into the learning method may lead to a significant decrease of the estimation error. In this paper, we present a family of convex penalty functions, which encode prior knowledge on the structure of the vector formed by the absolute values of the regression coefficients. This family subsumes the ℓ 1 norm and is flexible enough to include different models of sparsity patterns, which are of practical and theoretical importance. We establish the basic properties of these penalty functions and discuss some examples where they can be computed explicitly. Moreover, we present a convergent optimization algorithm for solving regularized least squares with these penalty functions. Numerical simulations highlight the benefit of structured sparsity and the advantage offered by our approach over the Lasso method and other related methods.</description><subject>Combinatorial analysis</subject><subject>Computational Mathematics and Numerical Analysis</subject><subject>Computational Science and Engineering</subject><subject>Learning</subject><subject>Mathematical analysis</subject><subject>Mathematical and Computational Biology</subject><subject>Mathematical Modeling and Industrial Mathematics</subject><subject>Mathematical models</subject><subject>Mathematics</subject><subject>Mathematics and Statistics</subject><subject>Norms</subject><subject>Penalty function</subject><subject>Regression</subject><subject>Vectors (mathematics)</subject><subject>Visualization</subject><issn>1019-7168</issn><issn>1572-9044</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2013</creationdate><recordtype>article</recordtype><recordid>eNp9kMFOwzAMhiMEEmPwANx64MAlEKdJUx_RBANpEhKCc5RmydSpW0fcHsbTk6mcOdmy_s-yP8ZuQTyAEOaRQCiluADgKJXmeMZmoI3kmOfnuReA3EBVX7Iroq0QAiujZ-zuI2zGzqX2JyQqYp8KGtLohzGFdUEHl6gdjtfsIrqOws1fnbOvl-fPxStfvS_fFk8r7ksJA48xKu-lwnXtPQLoMqJxCkWARqMCHYxWvqnQ16VXlc93N7GMtRMlNhhlOWf3095D6r_HQIPdteRD17l96EeyoCSiqLUxOQpT1KeeKIVoD6nduXS0IOzJiJ2M2GzEnoxYzIycGMrZ_SYku-3HtM8f_QP9AksfYtU</recordid><startdate>20130401</startdate><enddate>20130401</enddate><creator>Micchelli, Charles A.</creator><creator>Morales, Jean M.</creator><creator>Pontil, Massimiliano</creator><general>Springer US</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope></search><sort><creationdate>20130401</creationdate><title>Regularizers for structured sparsity</title><author>Micchelli, Charles A. ; Morales, Jean M. ; Pontil, Massimiliano</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c321t-fff4cc249d8cc91153f97a490e1b59415e754cb69c83c46c007bf3f8a039b9f23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2013</creationdate><topic>Combinatorial analysis</topic><topic>Computational Mathematics and Numerical Analysis</topic><topic>Computational Science and Engineering</topic><topic>Learning</topic><topic>Mathematical analysis</topic><topic>Mathematical and Computational Biology</topic><topic>Mathematical Modeling and Industrial Mathematics</topic><topic>Mathematical models</topic><topic>Mathematics</topic><topic>Mathematics and Statistics</topic><topic>Norms</topic><topic>Penalty function</topic><topic>Regression</topic><topic>Vectors (mathematics)</topic><topic>Visualization</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Micchelli, Charles A.</creatorcontrib><creatorcontrib>Morales, Jean M.</creatorcontrib><creatorcontrib>Pontil, Massimiliano</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts – Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>Advances in computational mathematics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Micchelli, Charles A.</au><au>Morales, Jean M.</au><au>Pontil, Massimiliano</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Regularizers for structured sparsity</atitle><jtitle>Advances in computational mathematics</jtitle><stitle>Adv Comput Math</stitle><date>2013-04-01</date><risdate>2013</risdate><volume>38</volume><issue>3</issue><spage>455</spage><epage>489</epage><pages>455-489</pages><issn>1019-7168</issn><eissn>1572-9044</eissn><abstract>We study the problem of learning a sparse linear regression vector under additional conditions on the structure of its sparsity pattern. This problem is relevant in machine learning, statistics and signal processing. It is well known that a linear regression can benefit from knowledge that the underlying regression vector is sparse. The combinatorial problem of selecting the nonzero components of this vector can be “relaxed” by regularizing the squared error with a convex penalty function like the ℓ 1 norm. However, in many applications, additional conditions on the structure of the regression vector and its sparsity pattern are available. Incorporating this information into the learning method may lead to a significant decrease of the estimation error. In this paper, we present a family of convex penalty functions, which encode prior knowledge on the structure of the vector formed by the absolute values of the regression coefficients. This family subsumes the ℓ 1 norm and is flexible enough to include different models of sparsity patterns, which are of practical and theoretical importance. We establish the basic properties of these penalty functions and discuss some examples where they can be computed explicitly. Moreover, we present a convergent optimization algorithm for solving regularized least squares with these penalty functions. Numerical simulations highlight the benefit of structured sparsity and the advantage offered by our approach over the Lasso method and other related methods.</abstract><cop>Boston</cop><pub>Springer US</pub><doi>10.1007/s10444-011-9245-9</doi><tpages>35</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1019-7168
ispartof Advances in computational mathematics, 2013-04, Vol.38 (3), p.455-489
issn 1019-7168
1572-9044
language eng
recordid cdi_proquest_miscellaneous_1429908577
source SpringerLink Journals - AutoHoldings
subjects Combinatorial analysis
Computational Mathematics and Numerical Analysis
Computational Science and Engineering
Learning
Mathematical analysis
Mathematical and Computational Biology
Mathematical Modeling and Industrial Mathematics
Mathematical models
Mathematics
Mathematics and Statistics
Norms
Penalty function
Regression
Vectors (mathematics)
Visualization
title Regularizers for structured sparsity
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T18%3A33%3A24IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Regularizers%20for%20structured%20sparsity&rft.jtitle=Advances%20in%20computational%20mathematics&rft.au=Micchelli,%20Charles%20A.&rft.date=2013-04-01&rft.volume=38&rft.issue=3&rft.spage=455&rft.epage=489&rft.pages=455-489&rft.issn=1019-7168&rft.eissn=1572-9044&rft_id=info:doi/10.1007/s10444-011-9245-9&rft_dat=%3Cproquest_cross%3E1429908577%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=1429908577&rft_id=info:pmid/&rfr_iscdi=true