How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm
Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior t...
Gespeichert in:
Veröffentlicht in: | The Journal of chemical physics 2024-12, Vol.161 (24) |
---|---|
1. Verfasser: | |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | 24 |
container_start_page | |
container_title | The Journal of chemical physics |
container_volume | 161 |
creator | Kang, Sungwoo |
description | Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities. |
doi_str_mv | 10.1063/5.0234287 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_crossref_primary_10_1063_5_0234287</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3148500726</sourcerecordid><originalsourceid>FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</originalsourceid><addsrcrecordid>eNp90M1O3DAUBWALgcp02gUvgCyxoZUytWPHP-zQqC1ISEioXUd2cjNjSOJgO4K-PW5n6IIFq7P57tHVQeiEkhUlgn2rVqRkvFTyAC0oUbqQQpNDtCCkpIUWRByjjzHeE0KoLPkHdMy0pExruUD2yj_hTTDTFo8wB9PnSE8-PGA3Jggm-cE1ePIJxuRMHzE8p6x9bxJc4DvfA_YdTlvAA8RoNlBMJkY3brDpNz64tB0-oaMuX8LnfS7R7x_ff62vipvbn9fry5uioRWThVG8a7jqmo5ZZazQmre0VI3tNJQMWkVpWUljibbaSsEt560iTGnKlYW2Y0t0vuudgn-cIaZ6cLGBvjcj-DnWLMOKEFmKTM_e0Hs_hzF_908JSqWgWX3ZqSb4GAN09RTcYMKfmpL67_B1Ve-Hz_Z03zjbAdr_8nXpDL7uQGxcMsn58Z22F4aoiyA</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3148611761</pqid></control><display><type>article</type><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><source>AIP Journals Complete</source><creator>Kang, Sungwoo</creator><creatorcontrib>Kang, Sungwoo</creatorcontrib><description>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</description><identifier>ISSN: 0021-9606</identifier><identifier>ISSN: 1089-7690</identifier><identifier>EISSN: 1089-7690</identifier><identifier>DOI: 10.1063/5.0234287</identifier><identifier>PMID: 39713997</identifier><identifier>CODEN: JCPSA6</identifier><language>eng</language><publisher>United States: American Institute of Physics</publisher><subject>Algorithms ; Density functional theory ; Extrapolation ; Graph neural networks ; IP (Internet Protocol) ; Machine learning ; Message passing ; Neural networks</subject><ispartof>The Journal of chemical physics, 2024-12, Vol.161 (24)</ispartof><rights>Author(s)</rights><rights>2024 Author(s). Published under an exclusive license by AIP Publishing.</rights><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</cites><orcidid>0000-0001-8177-8815</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://pubs.aip.org/jcp/article-lookup/doi/10.1063/5.0234287$$EHTML$$P50$$Gscitation$$H</linktohtml><link.rule.ids>314,776,780,790,4498,27901,27902,76126</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39713997$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Kang, Sungwoo</creatorcontrib><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><title>The Journal of chemical physics</title><addtitle>J Chem Phys</addtitle><description>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</description><subject>Algorithms</subject><subject>Density functional theory</subject><subject>Extrapolation</subject><subject>Graph neural networks</subject><subject>IP (Internet Protocol)</subject><subject>Machine learning</subject><subject>Message passing</subject><subject>Neural networks</subject><issn>0021-9606</issn><issn>1089-7690</issn><issn>1089-7690</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><recordid>eNp90M1O3DAUBWALgcp02gUvgCyxoZUytWPHP-zQqC1ISEioXUd2cjNjSOJgO4K-PW5n6IIFq7P57tHVQeiEkhUlgn2rVqRkvFTyAC0oUbqQQpNDtCCkpIUWRByjjzHeE0KoLPkHdMy0pExruUD2yj_hTTDTFo8wB9PnSE8-PGA3Jggm-cE1ePIJxuRMHzE8p6x9bxJc4DvfA_YdTlvAA8RoNlBMJkY3brDpNz64tB0-oaMuX8LnfS7R7x_ff62vipvbn9fry5uioRWThVG8a7jqmo5ZZazQmre0VI3tNJQMWkVpWUljibbaSsEt560iTGnKlYW2Y0t0vuudgn-cIaZ6cLGBvjcj-DnWLMOKEFmKTM_e0Hs_hzF_908JSqWgWX3ZqSb4GAN09RTcYMKfmpL67_B1Ve-Hz_Z03zjbAdr_8nXpDL7uQGxcMsn58Z22F4aoiyA</recordid><startdate>20241228</startdate><enddate>20241228</enddate><creator>Kang, Sungwoo</creator><general>American Institute of Physics</general><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>8FD</scope><scope>H8D</scope><scope>L7M</scope><scope>7X8</scope><orcidid>https://orcid.org/0000-0001-8177-8815</orcidid></search><sort><creationdate>20241228</creationdate><title>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</title><author>Kang, Sungwoo</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c1537-a84fc48fcf3b8ab6994d128cbf9e23ed811257ab09b9b764b44d80389148bedf3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Algorithms</topic><topic>Density functional theory</topic><topic>Extrapolation</topic><topic>Graph neural networks</topic><topic>IP (Internet Protocol)</topic><topic>Machine learning</topic><topic>Message passing</topic><topic>Neural networks</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Kang, Sungwoo</creatorcontrib><collection>PubMed</collection><collection>CrossRef</collection><collection>Technology Research Database</collection><collection>Aerospace Database</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>MEDLINE - Academic</collection><jtitle>The Journal of chemical physics</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Kang, Sungwoo</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm</atitle><jtitle>The Journal of chemical physics</jtitle><addtitle>J Chem Phys</addtitle><date>2024-12-28</date><risdate>2024</risdate><volume>161</volume><issue>24</issue><issn>0021-9606</issn><issn>1089-7690</issn><eissn>1089-7690</eissn><coden>JCPSA6</coden><abstract>Graph neural network interatomic potentials (GNN-IPs) are gaining significant attention due to their capability of learning from large datasets. Specifically, universal interatomic potentials based on GNN, usually trained with crystalline geometries, often exhibit remarkable extrapolative behavior toward untrained domains, such as surfaces and amorphous configurations. However, the origin of this extrapolation capability is not well understood. This work provides a theoretical explanation of how GNN-IPs extrapolate to untrained geometries. First, we demonstrate that GNN-IPs can capture non-local electrostatic interactions through the message-passing algorithm, as evidenced by tests on toy models and density-functional theory data. We find that GNN-IP models, SevenNet and MACE, accurately predict electrostatic forces in untrained domains, indicating that they have learned the exact functional form of the Coulomb interaction. Based on these results, we suggest that the ability to learn non-local electrostatic interactions, coupled with the embedding nature of GNN-IPs, explains their extrapolation ability. We find that the universal GNN-IP, SevenNet-0, effectively infers non-local Coulomb interactions in untrained domains but fails to extrapolate the non-local forces arising from the kinetic term, which supports the suggested theory. Finally, we address the impact of hyperparameters on the extrapolation performance of universal potentials, such as SevenNet-0 and MACE-MP-0, and discuss the limitations of the extrapolation capabilities.</abstract><cop>United States</cop><pub>American Institute of Physics</pub><pmid>39713997</pmid><doi>10.1063/5.0234287</doi><tpages>16</tpages><orcidid>https://orcid.org/0000-0001-8177-8815</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 0021-9606 |
ispartof | The Journal of chemical physics, 2024-12, Vol.161 (24) |
issn | 0021-9606 1089-7690 1089-7690 |
language | eng |
recordid | cdi_crossref_primary_10_1063_5_0234287 |
source | AIP Journals Complete |
subjects | Algorithms Density functional theory Extrapolation Graph neural networks IP (Internet Protocol) Machine learning Message passing Neural networks |
title | How graph neural network interatomic potentials extrapolate: Role of the message-passing algorithm |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T10%3A30%3A44IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=How%20graph%20neural%20network%20interatomic%20potentials%20extrapolate:%20Role%20of%20the%20message-passing%20algorithm&rft.jtitle=The%20Journal%20of%20chemical%20physics&rft.au=Kang,%20Sungwoo&rft.date=2024-12-28&rft.volume=161&rft.issue=24&rft.issn=0021-9606&rft.eissn=1089-7690&rft.coden=JCPSA6&rft_id=info:doi/10.1063/5.0234287&rft_dat=%3Cproquest_cross%3E3148500726%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3148611761&rft_id=info:pmid/39713997&rfr_iscdi=true |