How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm

Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of chemical physics 2024-12, Vol.161 (24)
1. Verfasser: Kang, Sungwoo
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 24
container_start_page
container_title The Journal of chemical physics
container_volume 161
creator Kang, Sungwoo
description Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.
doi_str_mv 10.1063/5.0234287
format Article
fullrecord <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1063_5_0234287</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3148500726</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</originalsourceid><addsrcrecordid>eNp90M1O3DAUBWALgcp02gUvgCyxoZUytWPHP-zQqC1ISEioXUd2cjNjSOJgO4K-PW5n6IIFq7P57tHVQeiEkhUlgn2rVqRkvFTyAC0oUbqQQpNDtCCkpIUWRByjjzHeE0KoLPkHdMy0pExruUD2yj_hTTDTFo8wB9PnSE8-PGA3Jggm-cE1ePIJxuRMHzE8p6x9bxJc4DvfA_YdTlvAA8RoNlBMJkY3brDpNz64tB0-oaMuX8LnfS7R7x_ff62vipvbn9fry5uioRWThVG8a7jqmo5ZZazQmre0VI3tNJQMWkVpWUljibbaSsEt560iTGnKlYW2Y0t0vuudgn-cIaZ6cLGBvjcj-DnWLMOKEFmKTM_e0Hs_hzF_908JSqWgWX3ZqSb4GAN09RTcYMKfmpL67_B1Ve-Hz_Z03zjbAdr_8nXpDL7uQGxcMsn58Z22F4aoiyA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3148611761</pqid></control><display><type>article</type><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><source>AIP Journals Complete</source><creator>Kang, Sungwoo</creator><creatorcontrib>Kang, Sungwoo</creatorcontrib><description>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</description><identifier>ISSN: 0021-9606</identifier><identifier>ISSN: 1089-7690</identifier><identifier>EISSN: 1089-7690</identifier><identifier>DOI: 10.1063/5.0234287</identifier><identifier>PMID: 39713997</identifier><identifier>CODEN: JCPSA6</identifier><language>eng</language><publisher>United States: American Institute of Physics</publisher><subject>Algorithms ; Density functional theory ; Extrapolation ; Graph neural networks ; IP (Internet Protocol) ; Machine learning ; Message passing ; Neural networks</subject><ispartof>The Journal of chemical physics, 2024-12, Vol.161 (24)</ispartof><rights>Author(s)</rights><rights>2024 Author(s). Published under an exclusive license by AIP Publishing.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</cites><orcidid>0000-0001-8177-8815</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://pubs.aip.org/jcp/article-lookup/doi/10.1063/5.0234287$$EHTML$$P50$$Gscitation$$H</linktohtml><link.rule.ids>314,776,780,790,4498,27901,27902,76126</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39713997$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Kang, Sungwoo</creatorcontrib><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><title>The Journal of chemical physics</title><addtitle>J Chem Phys</addtitle><description>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</description><subject>Algorithms</subject><subject>Density functional theory</subject><subject>Extrapolation</subject><subject>Graph neural networks</subject><subject>IP (Internet Protocol)</subject><subject>Machine learning</subject><subject>Message passing</subject><subject>Neural networks</subject><issn>0021-9606</issn><issn>1089-7690</issn><issn>1089-7690</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp90M1O3DAUBWALgcp02gUvgCyxoZUytWPHP-zQqC1ISEioXUd2cjNjSOJgO4K-PW5n6IIFq7P57tHVQeiEkhUlgn2rVqRkvFTyAC0oUbqQQpNDtCCkpIUWRByjjzHeE0KoLPkHdMy0pExruUD2yj_hTTDTFo8wB9PnSE8-PGA3Jggm-cE1ePIJxuRMHzE8p6x9bxJc4DvfA_YdTlvAA8RoNlBMJkY3brDpNz64tB0-oaMuX8LnfS7R7x_ff62vipvbn9fry5uioRWThVG8a7jqmo5ZZazQmre0VI3tNJQMWkVpWUljibbaSsEt560iTGnKlYW2Y0t0vuudgn-cIaZ6cLGBvjcj-DnWLMOKEFmKTM_e0Hs_hzF_908JSqWgWX3ZqSb4GAN09RTcYMKfmpL67_B1Ve-Hz_Z03zjbAdr_8nXpDL7uQGxcMsn58Z22F4aoiyA</recordid><startdate>20241228</startdate><enddate>20241228</enddate><creator>Kang, Sungwoo</creator><general>American Institute of Physics</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>8FD</scope><scope>H8D</scope><scope>L7M</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0001-8177-8815</orcidid></search><sort><creationdate>20241228</creationdate><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><author>Kang, Sungwoo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Density functional theory</topic><topic>Extrapolation</topic><topic>Graph neural networks</topic><topic>IP (Internet Protocol)</topic><topic>Machine learning</topic><topic>Message passing</topic><topic>Neural networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kang, Sungwoo</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Technology Research Database</collection><collection>Aerospace Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>MEDLINE - Academic</collection><jtitle>The Journal of chemical physics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kang, Sungwoo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</atitle><jtitle>The Journal of chemical physics</jtitle><addtitle>J Chem Phys</addtitle><date>2024-12-28</date><risdate>2024</risdate><volume>161</volume><issue>24</issue><issn>0021-9606</issn><issn>1089-7690</issn><eissn>1089-7690</eissn><coden>JCPSA6</coden><abstract>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</abstract><cop>United States</cop><pub>American Institute of Physics</pub><pmid>39713997</pmid><doi>10.1063/5.0234287</doi><tpages>16</tpages><orcidid>https://orcid.org/0000-0001-8177-8815</orcidid></addata></record>
fulltext fulltext
identifier ISSN: 0021-9606
ispartof The Journal of chemical physics, 2024-12, Vol.161 (24)
issn 0021-9606
1089-7690
1089-7690
language eng
recordid cdi_crossref_primary_10_1063_5_0234287
source AIP Journals Complete
subjects Algorithms
Density functional theory
Extrapolation
Graph neural networks
IP (Internet Protocol)
Machine learning
Message passing
Neural networks
title How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T10%3A30%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=How%20graph%20neural%20network%20interatomic%20potentials%20extrapolate:%20Role%20of%20the%20message-passing%20algorithm&rft.jtitle=The%20Journal%20of%20chemical%20physics&rft.au=Kang,%20Sungwoo&rft.date=2024-12-28&rft.volume=161&rft.issue=24&rft.issn=0021-9606&rft.eissn=1089-7690&rft.coden=JCPSA6&rft_id=info:doi/10.1063/5.0234287&rft_dat=%3Cproquest_cross%3E3148500726%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3148611761&rft_id=info:pmid/39713997&rfr_iscdi=true