Modeling Multi-Step Scientific Processes with Graph Transformer Networks
This work presents the use of graph learning for the prediction of multi-step experimental outcomes for applications across experimental research, including material science, chemistry, and biology. The viability of geometric learning for regression tasks was benchmarked against a collection of line...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | This work presents the use of graph learning for the prediction of multi-step
experimental outcomes for applications across experimental research, including
material science, chemistry, and biology. The viability of geometric learning
for regression tasks was benchmarked against a collection of linear models
through a combination of simulated and real-world data training studies. First,
a selection of five arbitrarily designed multi-step surrogate functions were
developed to reflect various features commonly found within experimental
processes. A graph transformer network outperformed all tested linear models in
scenarios that featured hidden interactions between process steps and sequence
dependent features, while retaining equivalent performance in sequence agnostic
scenarios. Then, a similar comparison was applied to real-world literature data
on algorithm guided colloidal atomic layer deposition. Using the complete
reaction sequence as training data, the graph neural network outperformed all
linear models in predicting the three spectral properties for most training set
sizes. Further implementation of graph neural networks and geometric
representation of scientific processes for the prediction of experiment
outcomes could lead to algorithm driven navigation of higher dimension
parameter spaces and efficient exploration of more dynamic systems. |
---|---|
DOI: | 10.48550/arxiv.2408.05425 |