Toward Physics-Informed Neural Networks for 3-D Multilayer Cloud Mask Reconstruction

Three-dimensional cloud retrievals are critical for understanding their impact on climate and other applications, such as aviation safety, weather prediction, and remote sensing. However, obtaining high-resolution and accurate vertical representation of clouds remains unsolved due to the limitations...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on geoscience and remote sensing 2023, Vol.61, p.1-14
Hauptverfasser: Wang, Yiding, Gong, Jie, Wu, Dong L., Ding, Leah
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Three-dimensional cloud retrievals are critical for understanding their impact on climate and other applications, such as aviation safety, weather prediction, and remote sensing. However, obtaining high-resolution and accurate vertical representation of clouds remains unsolved due to the limitations imposed by satellite instrumentation, viewing conditions, and the complexity of cloud dynamics. Cloud masks are essential for comprehending various cloud vertical properties, but deriving accurate 3-D cloud masks from 2-D satellite imagery data is a challenging task. To tackle these challenges, we introduce a physics-informed loss function for training deep learning models that can extend 2-D cloud images into 3-D cloud masks. The proposed loss, called CloudMask loss, is composed of two domain knowledge-informed loss terms: one for evaluating cloud position and thickness and the other for measuring the number of layers. By combining these loss terms, we improve the trainability of the deep learning models for more accurate and meaningful results. We apply the proposed loss function to different neural networks and demonstrate significant improvements in multilayer cloud mask reconstruction. Utilizing the same neural network architecture, our proposed loss outperforms standard binary cross-entropy (BCE) loss in terms of multilayer cloud classification accuracy, number of layers accuracy, and thickness mean absolute error (MAE). The proposed loss function can be readily integrated into various neural network architectures, resulting in substantial performance gains in 3-D cloud mask generation.
ISSN:0196-2892
1558-0644
DOI:10.1109/TGRS.2023.3329649