AstroM$^3$: A self-supervised multimodal model for astronomy
While machine-learned models are now routinely employed to facilitate astronomical inquiry, model inputs tend to be limited to a primary data source (namely images or time series) and, in the more advanced approaches, some metadata. Yet with the growing use of wide-field, multiplexed observational r...
Gespeichert in:
Hauptverfasser: | , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | While machine-learned models are now routinely employed to facilitate
astronomical inquiry, model inputs tend to be limited to a primary data source
(namely images or time series) and, in the more advanced approaches, some
metadata. Yet with the growing use of wide-field, multiplexed observational
resources, individual sources of interest often have a broad range of
observational modes available. Here we construct an astronomical multimodal
dataset and propose AstroM$^3$, a self-supervised pre-training approach that
enables a model to learn from multiple modalities simultaneously. Specifically,
we extend the CLIP (Contrastive Language-Image Pretraining) model to a trimodal
setting, allowing the integration of time-series photometry data, spectra, and
astrophysical metadata. In a fine-tuning supervised setting, our results
demonstrate that CLIP pre-training improves classification performance for
time-series photometry, where accuracy increases from 84.6% to 91.5%.
Furthermore, CLIP boosts classification accuracy by up to 12.6% when the
availability of labeled data is limited, showing the effectiveness of
leveraging larger corpora of unlabeled data. In addition to fine-tuned
classification, we can use the trained model in other downstream tasks that are
not explicitly contemplated during the construction of the self-supervised
model. In particular we show the efficacy of using the learned embeddings for
misclassifications identification, similarity search, and anomaly detection.
One surprising highlight is the "rediscovery" of Mira subtypes and two
Rotational variable subclasses using manifold learning and dimension reduction
algorithm. To our knowledge this is the first construction of an $n>2$ mode
model in astronomy. Extensions to $n>3$ modes is naturally anticipated with
this approach. |
---|---|
DOI: | 10.48550/arxiv.2411.08842 |