Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment

Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI),...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-06
1. Verfasser: Julius von Kügelgen
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Julius von Kügelgen
description Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.
doi_str_mv 10.48550/arxiv.2406.13371
format Article
fullrecord <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2406_13371</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3070859399</sourcerecordid><originalsourceid>FETCH-LOGICAL-a529-fbfad5317500ea200aea2390d11903d5af59fd40e363a22dc0a104340ff414a23</originalsourceid><addsrcrecordid>eNotkE9LAzEQxYMgWGo_gCcXvHbrJJN0G29S_FOoCFI9CcvUJJKyza7JbtVv79r2MsMMvze8eYxdcJjImVJwTfHH7yZCwnTCEQt-wgYCkeczKcQZG6W0AQAxLYRSOGDvC2ND652ndWWzOXWJquzFNtGmfk-tr0O2tBSDD5832WtIXWPjzidrxtlTV7U-f_P2e5xRMMf5Lux8rMO2l5-zU0dVsqNjH7LV_d1q_pgvnx8W89tlTkro3K0dGYW8UACWBAD1FTUYzjWgUeSUdkaCxSmSEOYDiINECc5JLnt0yC4PZ_evl030W4q_5X8E5T6Cnrg6EE2svzqb2nJTdzH0nkqEAmZKo9b4B5oUXvk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3070859399</pqid></control><display><type>article</type><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Julius von Kügelgen</creator><creatorcontrib>Julius von Kügelgen</creatorcontrib><description>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2406.13371</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Artificial intelligence ; Changing environments ; Cognitive tasks ; Complex systems ; Computer Science - Artificial Intelligence ; Computer Science - Learning ; Data collection ; Environment models ; Machine learning ; Representations ; Semantics ; Statistics - Machine Learning ; Unstructured data</subject><ispartof>arXiv.org, 2024-06</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,781,882,27906</link.rule.ids><backlink>$$Uhttps://doi.org/10.17863/CAM.106852$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.13371$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Julius von Kügelgen</creatorcontrib><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><title>arXiv.org</title><description>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</description><subject>Artificial intelligence</subject><subject>Changing environments</subject><subject>Cognitive tasks</subject><subject>Complex systems</subject><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Data collection</subject><subject>Environment models</subject><subject>Machine learning</subject><subject>Representations</subject><subject>Semantics</subject><subject>Statistics - Machine Learning</subject><subject>Unstructured data</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotkE9LAzEQxYMgWGo_gCcXvHbrJJN0G29S_FOoCFI9CcvUJJKyza7JbtVv79r2MsMMvze8eYxdcJjImVJwTfHH7yZCwnTCEQt-wgYCkeczKcQZG6W0AQAxLYRSOGDvC2ND652ndWWzOXWJquzFNtGmfk-tr0O2tBSDD5832WtIXWPjzidrxtlTV7U-f_P2e5xRMMf5Lux8rMO2l5-zU0dVsqNjH7LV_d1q_pgvnx8W89tlTkro3K0dGYW8UACWBAD1FTUYzjWgUeSUdkaCxSmSEOYDiINECc5JLnt0yC4PZ_evl030W4q_5X8E5T6Cnrg6EE2svzqb2nJTdzH0nkqEAmZKo9b4B5oUXvk</recordid><startdate>20240619</startdate><enddate>20240619</enddate><creator>Julius von Kügelgen</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20240619</creationdate><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><author>Julius von Kügelgen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a529-fbfad5317500ea200aea2390d11903d5af59fd40e363a22dc0a104340ff414a23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial intelligence</topic><topic>Changing environments</topic><topic>Cognitive tasks</topic><topic>Complex systems</topic><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Data collection</topic><topic>Environment models</topic><topic>Machine learning</topic><topic>Representations</topic><topic>Semantics</topic><topic>Statistics - Machine Learning</topic><topic>Unstructured data</topic><toplevel>online_resources</toplevel><creatorcontrib>Julius von Kügelgen</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Julius von Kügelgen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</atitle><jtitle>arXiv.org</jtitle><date>2024-06-19</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2406.13371</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-06
issn 2331-8422
language eng
recordid cdi_arxiv_primary_2406_13371
source arXiv.org; Free E- Journals
subjects Artificial intelligence
Changing environments
Cognitive tasks
Complex systems
Computer Science - Artificial Intelligence
Computer Science - Learning
Data collection
Environment models
Machine learning
Representations
Semantics
Statistics - Machine Learning
Unstructured data
title Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T14%3A09%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Identifiable%20Causal%20Representation%20Learning:%20Unsupervised,%20Multi-View,%20and%20Multi-Environment&rft.jtitle=arXiv.org&rft.au=Julius%20von%20K%C3%BCgelgen&rft.date=2024-06-19&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2406.13371&rft_dat=%3Cproquest_arxiv%3E3070859399%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3070859399&rft_id=info:pmid/&rfr_iscdi=true