SSL4EO-S12: A Large-Scale Multi-Modal, Multi-Temporal Dataset for Self-Supervised Learning in Earth Observation

Self-supervised pre-training bears potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pre-training in the field is based on ImageNet or medium-size, labeled remote sensing (RS) datasets. In this paper, we sha...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Wang, Yi Wang, Ait Ali Braham, Nassim Ait Ali Braham, Xiong, Zhitong Xiong, Liu, Chenying Liu, Albrecht, Conrad M. Albrecht, Zhu, Xiao Xiang Zhu
Format: Dataset
Sprache:eng
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Self-supervised pre-training bears potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pre-training in the field is based on ImageNet or medium-size, labeled remote sensing (RS) datasets. In this paper, we share an unlabeled dataset SSL4EO-S12: Self-Supervised Learning for Earth Observation - Sentinel-1/2, to assemble a large-scale, global, multimodal, and multi-seasonal corpus of satellite imagery. We demonstrate SSL4EO-S12 to succeed in self-supervised pre-training for a set of representative methods: MoCo-v2, DINO, MAE and data2vec, and multiple downstream applications including scene classification, semantic segmentation and change detection. Our benchmark results prove the effectiveness of SSL4EO-S12 compared to existing datasets. The dataset, related source code, and pre-trained models are available at https://github.com/zhu-xlab/SSL4EO-S12.
DOI:10.21227/gkc0-3b82