Machine learning techniques to construct patched analog ensembles for data assimilation
•Domains need to be partitioned when constructing analogs for geophysical models.•Patches make the training of machine learning models more robust.•The use of patches makes in data assimilation can be implemented in parallel.•General autoencoders with an affine transformation in the latent space can...
Gespeichert in:
Veröffentlicht in: | Journal of computational physics 2021-10, Vol.443, p.110532, Article 110532 |
---|---|
Hauptverfasser: | , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | •Domains need to be partitioned when constructing analogs for geophysical models.•Patches make the training of machine learning models more robust.•The use of patches makes in data assimilation can be implemented in parallel.•General autoencoders with an affine transformation in the latent space can be used.•Patched constructed analogs can approximate ensemble members within DA methods.
Using generative models from the machine learning literature to create artificial ensemble members for use within data assimilation schemes has been introduced in Grooms (2021) [1] as constructed analog ensemble optimal interpolation (cAnEnOI). Specifically, we study general and variational autoencoders for the machine learning component of this method, and combine the ideas of constructed analogs and ensemble optimal interpolation in the data assimilation piece. To extend the scalability of cAnEnOI for use in data assimilation on complex dynamical models, we propose using patching schemes to divide the global spatial domain into digestible chunks. Using patches makes training the generative models possible and has the added benefit of being able to exploit parallelism during the generative step. Testing this new algorithm on a 1D toy model, we find that larger patch sizes make it harder to train an accurate generative model (i.e. a model whose reconstruction error is small), while conversely the data assimilation performance improves at larger patch sizes. There is thus a sweet spot where the patch size is large enough to enable good data assimilation performance, but not so large that it becomes difficult to train an accurate generative model. In our tests the new patched cAnEnOI method outperforms the original (unpatched) cAnEnOI, as well as the ensemble square root filter results from [1]. |
---|---|
ISSN: | 0021-9991 1090-2716 |
DOI: | 10.1016/j.jcp.2021.110532 |