SpectralEarth: Training Hyperspectral Foundation Models at Scale

Foundation models have triggered a paradigm shift in computer vision and are increasingly being adopted in remote sensing, particularly for multispectral imagery. Yet, their potential in hyperspectral imaging (HSI) remains untapped due to the absence of comprehensive and globally representative hype...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Braham, Nassim Ait Ali, Albrecht, Conrad M, Mairal, Julien, Chanussot, Jocelyn, Wang, Yi, Zhu, Xiao Xiang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Braham, Nassim Ait Ali
Albrecht, Conrad M
Mairal, Julien
Chanussot, Jocelyn
Wang, Yi
Zhu, Xiao Xiang
description Foundation models have triggered a paradigm shift in computer vision and are increasingly being adopted in remote sensing, particularly for multispectral imagery. Yet, their potential in hyperspectral imaging (HSI) remains untapped due to the absence of comprehensive and globally representative hyperspectral datasets. To close this gap, we introduce SpectralEarth, a large-scale multi-temporal dataset designed to pretrain hyperspectral foundation models leveraging data from the Environmental Mapping and Analysis Program (EnMAP). SpectralEarth comprises 538,974 image patches covering 415,153 unique locations from more than 11,636 globally distributed EnMAP scenes spanning two years of archive. Additionally, 17.5% of these locations include multiple timestamps, enabling multi-temporal HSI analysis. Utilizing state-of-the-art self-supervised learning (SSL) algorithms, we pretrain a series of foundation models on SpectralEarth. We integrate a spectral adapter into classical vision backbones to accommodate the unique characteristics of HSI. In tandem, we construct four downstream datasets for land-cover and crop-type mapping, providing benchmarks for model evaluation. Experimental results support the versatility of our models, showcasing their generalizability across different tasks and sensors. We also highlight computational efficiency during model fine-tuning. The dataset, models, and source code will be made publicly available.
doi_str_mv 10.48550/arxiv.2408.08447
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2408_08447</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2408_08447</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2408_084473</originalsourceid><addsrcrecordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw0DOwMDEx52RwCC5ITS4pSsxxTSwqybBSCClKzMzLzEtX8KgsSC0qhkoquOWX5qUklmTm5yn45qek5hQrJJYoBCcn5qTyMLCmJeYUp_JCaW4GeTfXEGcPXbBd8QVFmbmJRZXxIDvjwXYaE1YBABPLNuk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>SpectralEarth: Training Hyperspectral Foundation Models at Scale</title><source>arXiv.org</source><creator>Braham, Nassim Ait Ali ; Albrecht, Conrad M ; Mairal, Julien ; Chanussot, Jocelyn ; Wang, Yi ; Zhu, Xiao Xiang</creator><creatorcontrib>Braham, Nassim Ait Ali ; Albrecht, Conrad M ; Mairal, Julien ; Chanussot, Jocelyn ; Wang, Yi ; Zhu, Xiao Xiang</creatorcontrib><description>Foundation models have triggered a paradigm shift in computer vision and are increasingly being adopted in remote sensing, particularly for multispectral imagery. Yet, their potential in hyperspectral imaging (HSI) remains untapped due to the absence of comprehensive and globally representative hyperspectral datasets. To close this gap, we introduce SpectralEarth, a large-scale multi-temporal dataset designed to pretrain hyperspectral foundation models leveraging data from the Environmental Mapping and Analysis Program (EnMAP). SpectralEarth comprises 538,974 image patches covering 415,153 unique locations from more than 11,636 globally distributed EnMAP scenes spanning two years of archive. Additionally, 17.5% of these locations include multiple timestamps, enabling multi-temporal HSI analysis. Utilizing state-of-the-art self-supervised learning (SSL) algorithms, we pretrain a series of foundation models on SpectralEarth. We integrate a spectral adapter into classical vision backbones to accommodate the unique characteristics of HSI. In tandem, we construct four downstream datasets for land-cover and crop-type mapping, providing benchmarks for model evaluation. Experimental results support the versatility of our models, showcasing their generalizability across different tasks and sensors. We also highlight computational efficiency during model fine-tuning. The dataset, models, and source code will be made publicly available.</description><identifier>DOI: 10.48550/arxiv.2408.08447</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Vision and Pattern Recognition</subject><creationdate>2024-08</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2408.08447$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2408.08447$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Braham, Nassim Ait Ali</creatorcontrib><creatorcontrib>Albrecht, Conrad M</creatorcontrib><creatorcontrib>Mairal, Julien</creatorcontrib><creatorcontrib>Chanussot, Jocelyn</creatorcontrib><creatorcontrib>Wang, Yi</creatorcontrib><creatorcontrib>Zhu, Xiao Xiang</creatorcontrib><title>SpectralEarth: Training Hyperspectral Foundation Models at Scale</title><description>Foundation models have triggered a paradigm shift in computer vision and are increasingly being adopted in remote sensing, particularly for multispectral imagery. Yet, their potential in hyperspectral imaging (HSI) remains untapped due to the absence of comprehensive and globally representative hyperspectral datasets. To close this gap, we introduce SpectralEarth, a large-scale multi-temporal dataset designed to pretrain hyperspectral foundation models leveraging data from the Environmental Mapping and Analysis Program (EnMAP). SpectralEarth comprises 538,974 image patches covering 415,153 unique locations from more than 11,636 globally distributed EnMAP scenes spanning two years of archive. Additionally, 17.5% of these locations include multiple timestamps, enabling multi-temporal HSI analysis. Utilizing state-of-the-art self-supervised learning (SSL) algorithms, we pretrain a series of foundation models on SpectralEarth. We integrate a spectral adapter into classical vision backbones to accommodate the unique characteristics of HSI. In tandem, we construct four downstream datasets for land-cover and crop-type mapping, providing benchmarks for model evaluation. Experimental results support the versatility of our models, showcasing their generalizability across different tasks and sensors. We also highlight computational efficiency during model fine-tuning. The dataset, models, and source code will be made publicly available.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpjYJA0NNAzsTA1NdBPLKrILNMzMjGw0DOwMDEx52RwCC5ITS4pSsxxTSwqybBSCClKzMzLzEtX8KgsSC0qhkoquOWX5qUklmTm5yn45qek5hQrJJYoBCcn5qTyMLCmJeYUp_JCaW4GeTfXEGcPXbBd8QVFmbmJRZXxIDvjwXYaE1YBABPLNuk</recordid><startdate>20240815</startdate><enddate>20240815</enddate><creator>Braham, Nassim Ait Ali</creator><creator>Albrecht, Conrad M</creator><creator>Mairal, Julien</creator><creator>Chanussot, Jocelyn</creator><creator>Wang, Yi</creator><creator>Zhu, Xiao Xiang</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240815</creationdate><title>SpectralEarth: Training Hyperspectral Foundation Models at Scale</title><author>Braham, Nassim Ait Ali ; Albrecht, Conrad M ; Mairal, Julien ; Chanussot, Jocelyn ; Wang, Yi ; Zhu, Xiao Xiang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2408_084473</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><toplevel>online_resources</toplevel><creatorcontrib>Braham, Nassim Ait Ali</creatorcontrib><creatorcontrib>Albrecht, Conrad M</creatorcontrib><creatorcontrib>Mairal, Julien</creatorcontrib><creatorcontrib>Chanussot, Jocelyn</creatorcontrib><creatorcontrib>Wang, Yi</creatorcontrib><creatorcontrib>Zhu, Xiao Xiang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Braham, Nassim Ait Ali</au><au>Albrecht, Conrad M</au><au>Mairal, Julien</au><au>Chanussot, Jocelyn</au><au>Wang, Yi</au><au>Zhu, Xiao Xiang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>SpectralEarth: Training Hyperspectral Foundation Models at Scale</atitle><date>2024-08-15</date><risdate>2024</risdate><abstract>Foundation models have triggered a paradigm shift in computer vision and are increasingly being adopted in remote sensing, particularly for multispectral imagery. Yet, their potential in hyperspectral imaging (HSI) remains untapped due to the absence of comprehensive and globally representative hyperspectral datasets. To close this gap, we introduce SpectralEarth, a large-scale multi-temporal dataset designed to pretrain hyperspectral foundation models leveraging data from the Environmental Mapping and Analysis Program (EnMAP). SpectralEarth comprises 538,974 image patches covering 415,153 unique locations from more than 11,636 globally distributed EnMAP scenes spanning two years of archive. Additionally, 17.5% of these locations include multiple timestamps, enabling multi-temporal HSI analysis. Utilizing state-of-the-art self-supervised learning (SSL) algorithms, we pretrain a series of foundation models on SpectralEarth. We integrate a spectral adapter into classical vision backbones to accommodate the unique characteristics of HSI. In tandem, we construct four downstream datasets for land-cover and crop-type mapping, providing benchmarks for model evaluation. Experimental results support the versatility of our models, showcasing their generalizability across different tasks and sensors. We also highlight computational efficiency during model fine-tuning. The dataset, models, and source code will be made publicly available.</abstract><doi>10.48550/arxiv.2408.08447</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2408.08447
ispartof
issn
language eng
recordid cdi_arxiv_primary_2408_08447
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Vision and Pattern Recognition
title SpectralEarth: Training Hyperspectral Foundation Models at Scale
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T10%3A55%3A14IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=SpectralEarth:%20Training%20Hyperspectral%20Foundation%20Models%20at%20Scale&rft.au=Braham,%20Nassim%20Ait%20Ali&rft.date=2024-08-15&rft_id=info:doi/10.48550/arxiv.2408.08447&rft_dat=%3Carxiv_GOX%3E2408_08447%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true