SSL4EO-S12: A Large-Scale Multi-Modal, Multi-Temporal Dataset for Self-Supervised Learning in Earth Observation
Self-supervised pre-training bears potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pre-training in the field is based on ImageNet or medium-size, labeled remote sensing (RS) datasets. In this paper, we sha...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Dataset |
Sprache: | eng |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Self-supervised pre-training bears potential to generate expressive representations from large-scale Earth observation (EO) data without human annotation. However, most existing pre-training in the field is based on ImageNet or medium-size, labeled remote sensing (RS) datasets. In this paper, we share an unlabeled dataset SSL4EO-S12: Self-Supervised Learning for Earth Observation - Sentinel-1/2, to assemble a large-scale, global, multimodal, and multi-seasonal corpus of satellite imagery. We demonstrate SSL4EO-S12 to succeed in self-supervised pre-training for a set of representative methods: MoCo-v2, DINO, MAE and data2vec, and multiple downstream applications including scene classification, semantic segmentation and change detection. Our benchmark results prove the effectiveness of SSL4EO-S12 compared to existing datasets. The dataset, related source code, and pre-trained models are available at https://github.com/zhu-xlab/SSL4EO-S12. |
---|---|
DOI: | 10.21227/gkc0-3b82 |