Symplectic learning for Hamiltonian neural networks
Machine learning methods are widely used in the natural sciences to model and predict physical systems from observation data. Yet, they are often used as poorly understood “black boxes,” disregarding existing mathematical structure and invariants of the problem. Recently, the proposal of Hamiltonian...
Gespeichert in:
Veröffentlicht in: | Journal of computational physics 2023-12, Vol.494, p.112495, Article 112495 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Machine learning methods are widely used in the natural sciences to model and predict physical systems from observation data. Yet, they are often used as poorly understood “black boxes,” disregarding existing mathematical structure and invariants of the problem. Recently, the proposal of Hamiltonian Neural Networks (HNNs) took a first step towards a unified “gray box” approach, using physical insight to improve performance for Hamiltonian systems. In this paper, we explore a significantly improved training method for HNNs, exploiting the symplectic structure of Hamiltonian systems with a different loss function. This frees the loss from an artificial lower bound. We mathematically guarantee the existence of an exact Hamiltonian function which the HNN can learn. This allows us to prove and numerically analyze the errors made by HNNs which, in turn, renders them fully explainable. Finally, we present a novel post-training correction to obtain the true Hamiltonian only from discretized observation data, up to an arbitrary order.
•The loss of Hamiltonian Neural Networks suffers from an artificial lower bound.•Using a symplectic integrator during training significantly improves performance.•Rigorous mathematical analysis yields an exact handle on the discretization error.•The learnt Hamiltonian can be corrected post-training, to any arbitrary order. |
---|---|
ISSN: | 0021-9991 1090-2716 |
DOI: | 10.1016/j.jcp.2023.112495 |