AstroMLab 3: Achieving GPT-4o Level Performance in Astronomy with a Specialized 8B-Parameter Large Language Model
AstroSage-Llama-3.1-8B is a domain-specialized natural-language AI assistant tailored for research in astronomy, astrophysics, and cosmology. Trained on the complete collection of astronomy-related arXiv papers from 2007-2024 along with millions of synthetically-generated question-answer pairs and o...
Gespeichert in:
Hauptverfasser: | , , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | AstroSage-Llama-3.1-8B is a domain-specialized natural-language AI assistant
tailored for research in astronomy, astrophysics, and cosmology. Trained on the
complete collection of astronomy-related arXiv papers from 2007-2024 along with
millions of synthetically-generated question-answer pairs and other
astronomical literature, AstroSage-Llama-3.1-8B demonstrates remarkable
proficiency on a wide range of questions. AstroSage-Llama-3.1-8B scores 80.9%
on the AstroMLab-1 benchmark, greatly outperforming all models -- proprietary
and open-weight -- in the 8-billion parameter class, and performing on par with
GPT-4o. This achievement demonstrates the potential of domain specialization in
AI, suggesting that focused training can yield capabilities exceeding those of
much larger, general-purpose models. AstroSage-Llama-3.1-8B is freely
available, enabling widespread access to advanced AI capabilities for
astronomical education and research. |
---|---|
DOI: | 10.48550/arxiv.2411.09012 |