AstroSage-8B / README.md
tingyuansen's picture
Update README.md
c3eb13e verified
|
raw
history blame
3.8 kB
---
language:
- en
tags:
- physics
- astronomy
- astrophysics
- cosmology
license:
- llama3.1
base_model:
- meta-llama/Meta-Llama-3.1-8B
library_name: transformers
---
# AstroSage-Llama-3.1-8B
<INSERT PAPER LINK HERE>
AstroSage-Llama-3.1-8B is a domain-specialized natural-language AI assistant
tailored for research in astronomy, astrophysics, and cosmology. Trained on the
complete collection of astronomy-related arXiv papers from 2007-2024 along with
millions of synthetically-generated question-answer pairs and other
astronomical literature, AstroSage-Llama-3.1-8B demonstrates excellent
proficiency on a wide range of questions. AstroSage-Llama-3.1-8B scores 80.9%
on the AstroMLab-1 benchmark, greatly outperforming all models---proprietary
and open-weight---in the 8-billion parameter class, and performing on par with
GPT-4o. This achievement demonstrates the potential of domain specialization in
AI, suggesting that focused training can yield capabilities exceeding those of
much larger, general-purpose models. AstroSage-Llama-3.1-8B is freely
available, enabling widespread access to advanced AI capabilities for
astronomical education and research.
## Model Details
- **Model Type**: Domain-specialized LLM
- **Base Model**: Meta-Llama-3.1-8B
- **Parameters**: 8 billion
- **Training Focus**: Astronomy, Astrophysics, Cosmology, and Astronomical Instrumentation
- **License**: Llama 3.1 Community License
- **Development Process**:
1. Continued Pre-training (CPT) on astronomical literature
2. Supervised Fine-tuning (SFT) on QA pairs and instruction sets
3. Model merging with Meta-Llama-3.1-8B-Instruct (75% CPT+SFT / 25% Meta-Instruct)
## Performance
- **AstroMLab-1 Benchmark**: 80.9% accuracy
- Outperforms all 8B parameter models
- Comparable to GPT-4o (80.4%)
- ~1000x more cost-effective than proprietary models
- 8 percentage-point improvement over base Llama-3.1-8b model on Astronomy Q&A benchmark
![image/png](https://cdn-uploads.huggingface.co/production/uploads/643f1ddce2ea47d170103537/9cXdGN7xf4vDag0_bgJcL.png)
- **General Capabilities**: Maintains strong performance on standard benchmarks
- IF-EVAL: 41.4%
- BBH: 52.9%
- MATH: 8.4%
- GPQA: 31.2%
- MUSR: 38.9%
- MMLU-PRO: 34.6%
## Training Data
- **Continued Pre-training**:
- ~250,000 arXiv preprints (2007-2024) from astro-ph and gr-qc
- Astronomy-related Wikipedia articles
- Selected astronomy textbooks
- Total: 3.3 billion tokens, 19.9 GB plaintext
- **Supervised Fine-tuning**:
- 8.8 million curated QA pairs
- Filtered Infinity-Instruct-7M dataset
- Paper summaries and metadata
- Total: 2.0 billion tokens, 9.8 GB plaintext
## Intended Use
- Curiosity-driven question answering
- Brainstorming new ideas
- Astronomical research assistance
- Educational support in astronomy
- Literature review and summarization
- Scientific explanation of concepts
## Limitations
- Training data cutoff: January 2024
- As with all LLMs, hallucinations are possible
- Limited by 8B parameter size for complex reasoning
- Paper metadata not perfectly memorized
- Performance primarily validated on multiple-choice questions
- Primarily trained for use in English
## Ethical Considerations
- Should not be used as sole source for critical research decisions
- Output should be verified against primary sources
- May reflect biases present in astronomical literature
## Technical Specifications
- Architecture: Based on Meta-Llama 3.1
- Training Infrastructure: ORNL OLCF Frontier
- Hosting: Hugging Face Hub (AstroMLab/AstroSage-8B)
## Citation and Contact
- Contract: Corresponding author Tijmen de Haan, email: tijmen dot dehaan at gmail dot com and AstroMLab astromachinelearninglab at gmail dot com
- Please cite the AstroMLab 3 paper when referencing to this model.