tingyuansen commited on
Commit
ed8cab4
1 Parent(s): 177cd69

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -35,7 +35,7 @@ AstroLLaMA-3-8B-Base_Summary is a specialized base language model for astronomy,
35
  - Cosine decay schedule for learning rate reduction
36
  - Training duration: 1 epoch
37
  - **Primary Use**: Next token prediction for astronomy-related text generation and analysis
38
- - **Reference**: Pan et al. 2024 [Link to be added]
39
 
40
  ## Generating text from a prompt
41
 
@@ -79,6 +79,7 @@ Here's a performance comparison chart based upon the astronomical benchmarking Q
79
 
80
  | Model | Score (%) |
81
  |-------|-----------|
 
82
  | LLaMA-3.1-8B | 73.7 |
83
  | LLaMA-3-8B | 72.9 |
84
  | **<span style="color:green">AstroLLaMA-3-8B-Base_Summary (AstroMLab)</span>** | **<span style="color:green">72.3</span>** |
@@ -92,7 +93,7 @@ Here's a performance comparison chart based upon the astronomical benchmarking Q
92
 
93
  As shown, AstroLLaMA-3-8B-Base_Summary performs competitively, nearly matching the performance of the base LLaMA-3.1-8B model and outperforming the AIC version. This improvement demonstrates the importance of information density in the training data.
94
 
95
- Notably, the instruct version of this model shows even more significant improvements, highlighting the effectiveness of the summarization approach in capturing and retaining key astronomical concepts. For detailed performance analysis of the instruct version, please refer to Pan et al. 2024.
96
 
97
  While AstroLLaMA-3-8B performs competitively among models in its class, it does not surpass the performance of the base LLaMA-3-8B model. This underscores the challenges in developing specialized models and the need for more diverse and comprehensive training data.
98
 
 
35
  - Cosine decay schedule for learning rate reduction
36
  - Training duration: 1 epoch
37
  - **Primary Use**: Next token prediction for astronomy-related text generation and analysis
38
+ - **Reference**: [Pan et al. 2024](https://arxiv.org/abs/2409.19750)
39
 
40
  ## Generating text from a prompt
41
 
 
79
 
80
  | Model | Score (%) |
81
  |-------|-----------|
82
+ | **AstroSage-LLaMA-3.1-8B (AstroMLab)** | **80.9** |
83
  | LLaMA-3.1-8B | 73.7 |
84
  | LLaMA-3-8B | 72.9 |
85
  | **<span style="color:green">AstroLLaMA-3-8B-Base_Summary (AstroMLab)</span>** | **<span style="color:green">72.3</span>** |
 
93
 
94
  As shown, AstroLLaMA-3-8B-Base_Summary performs competitively, nearly matching the performance of the base LLaMA-3.1-8B model and outperforming the AIC version. This improvement demonstrates the importance of information density in the training data.
95
 
96
+ Notably, the instruct version of this model shows even more significant improvements, highlighting the effectiveness of the summarization approach in capturing and retaining key astronomical concepts. For detailed performance analysis of the instruct version, please refer to [Pan et al. 2024](https://arxiv.org/abs/2409.19750).
97
 
98
  While AstroLLaMA-3-8B performs competitively among models in its class, it does not surpass the performance of the base LLaMA-3-8B model. This underscores the challenges in developing specialized models and the need for more diverse and comprehensive training data.
99