Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment
Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI),...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2024-06 |
---|---|
1. Verfasser: | |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Julius von Kügelgen |
description | Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods. |
doi_str_mv | 10.48550/arxiv.2406.13371 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2406_13371</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3070859399</sourcerecordid><originalsourceid>FETCH-LOGICAL-a529-fbfad5317500ea200aea2390d11903d5af59fd40e363a22dc0a104340ff414a23</originalsourceid><addsrcrecordid>eNotkE9LAzEQxYMgWGo_gCcXvHbrJJN0G29S_FOoCFI9CcvUJJKyza7JbtVv79r2MsMMvze8eYxdcJjImVJwTfHH7yZCwnTCEQt-wgYCkeczKcQZG6W0AQAxLYRSOGDvC2ND652ndWWzOXWJquzFNtGmfk-tr0O2tBSDD5832WtIXWPjzidrxtlTV7U-f_P2e5xRMMf5Lux8rMO2l5-zU0dVsqNjH7LV_d1q_pgvnx8W89tlTkro3K0dGYW8UACWBAD1FTUYzjWgUeSUdkaCxSmSEOYDiINECc5JLnt0yC4PZ_evl030W4q_5X8E5T6Cnrg6EE2svzqb2nJTdzH0nkqEAmZKo9b4B5oUXvk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3070859399</pqid></control><display><type>article</type><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Julius von Kügelgen</creator><creatorcontrib>Julius von Kügelgen</creatorcontrib><description>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2406.13371</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Artificial intelligence ; Changing environments ; Cognitive tasks ; Complex systems ; Computer Science - Artificial Intelligence ; Computer Science - Learning ; Data collection ; Environment models ; Machine learning ; Representations ; Semantics ; Statistics - Machine Learning ; Unstructured data</subject><ispartof>arXiv.org, 2024-06</ispartof><rights>2024. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by-nc-nd/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,781,882,27906</link.rule.ids><backlink>$$Uhttps://doi.org/10.17863/CAM.106852$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.13371$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Julius von Kügelgen</creatorcontrib><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><title>arXiv.org</title><description>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</description><subject>Artificial intelligence</subject><subject>Changing environments</subject><subject>Cognitive tasks</subject><subject>Complex systems</subject><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Learning</subject><subject>Data collection</subject><subject>Environment models</subject><subject>Machine learning</subject><subject>Representations</subject><subject>Semantics</subject><subject>Statistics - Machine Learning</subject><subject>Unstructured data</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotkE9LAzEQxYMgWGo_gCcXvHbrJJN0G29S_FOoCFI9CcvUJJKyza7JbtVv79r2MsMMvze8eYxdcJjImVJwTfHH7yZCwnTCEQt-wgYCkeczKcQZG6W0AQAxLYRSOGDvC2ND652ndWWzOXWJquzFNtGmfk-tr0O2tBSDD5832WtIXWPjzidrxtlTV7U-f_P2e5xRMMf5Lux8rMO2l5-zU0dVsqNjH7LV_d1q_pgvnx8W89tlTkro3K0dGYW8UACWBAD1FTUYzjWgUeSUdkaCxSmSEOYDiINECc5JLnt0yC4PZ_evl030W4q_5X8E5T6Cnrg6EE2svzqb2nJTdzH0nkqEAmZKo9b4B5oUXvk</recordid><startdate>20240619</startdate><enddate>20240619</enddate><creator>Julius von Kügelgen</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20240619</creationdate><title>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</title><author>Julius von Kügelgen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a529-fbfad5317500ea200aea2390d11903d5af59fd40e363a22dc0a104340ff414a23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Artificial intelligence</topic><topic>Changing environments</topic><topic>Cognitive tasks</topic><topic>Complex systems</topic><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Learning</topic><topic>Data collection</topic><topic>Environment models</topic><topic>Machine learning</topic><topic>Representations</topic><topic>Semantics</topic><topic>Statistics - Machine Learning</topic><topic>Unstructured data</topic><toplevel>online_resources</toplevel><creatorcontrib>Julius von Kügelgen</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Julius von Kügelgen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment</atitle><jtitle>arXiv.org</jtitle><date>2024-06-19</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Causal models provide rich descriptions of complex systems as sets of mechanisms by which each variable is influenced by its direct causes. They support reasoning about manipulating parts of the system and thus hold promise for addressing some of the open challenges of artificial intelligence (AI), such as planning, transferring knowledge in changing environments, or robustness to distribution shifts. However, a key obstacle to more widespread use of causal models in AI is the requirement that the relevant variables be specified a priori, which is typically not the case for the high-dimensional, unstructured data processed by modern AI systems. At the same time, machine learning (ML) has proven quite successful at automatically extracting useful and compact representations of such complex data. Causal representation learning (CRL) aims to combine the core strengths of ML and causality by learning representations in the form of latent variables endowed with causal model semantics. In this thesis, we study and present new results for different CRL settings. A central theme is the question of identifiability: Given infinite data, when are representations satisfying the same learning objective guaranteed to be equivalent? This is an important prerequisite for CRL, as it formally characterises if and when a learning task is, at least in principle, feasible. Since learning causal models, even without a representation learning component, is notoriously difficult, we require additional assumptions on the model class or rich data beyond the classical i.i.d. setting. By partially characterising identifiability for different settings, this thesis investigates what is possible for CRL without direct supervision, and thus contributes to its theoretical foundations. Ideally, the developed insights can help inform data collection practices or inspire the design of new practical estimation methods.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2406.13371</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2024-06 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2406_13371 |
source | arXiv.org; Free E- Journals |
subjects | Artificial intelligence Changing environments Cognitive tasks Complex systems Computer Science - Artificial Intelligence Computer Science - Learning Data collection Environment models Machine learning Representations Semantics Statistics - Machine Learning Unstructured data |
title | Identifiable Causal Representation Learning: Unsupervised, Multi-View, and Multi-Environment |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T14%3A09%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Identifiable%20Causal%20Representation%20Learning:%20Unsupervised,%20Multi-View,%20and%20Multi-Environment&rft.jtitle=arXiv.org&rft.au=Julius%20von%20K%C3%BCgelgen&rft.date=2024-06-19&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2406.13371&rft_dat=%3Cproquest_arxiv%3E3070859399%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3070859399&rft_id=info:pmid/&rfr_iscdi=true |