Weyaxi commited on
Commit
c68a8b7
1 Parent(s): 1b36e5f

good readme

Browse files
Files changed (1) hide show
  1. README.md +110 -35
README.md CHANGED
@@ -1,15 +1,66 @@
1
  ---
 
 
2
  license: other
3
  tags:
4
  - axolotl
5
  - generated_from_trainer
 
 
 
 
 
 
 
 
 
 
 
 
 
6
  base_model: Qwen/Qwen1.5-32B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  ---
 
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
11
 
12
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
13
  <details><summary>See axolotl config</summary>
14
 
15
  axolotl version: `0.4.0`
@@ -121,54 +172,78 @@ special_tokens:
121
  unk_token: "<unk>"
122
  tokens:
123
  - "<|im_start|>"
 
 
 
 
 
 
 
 
124
 
 
 
 
 
 
 
 
125
  ```
126
 
127
- </details><br>
 
128
 
129
- # Einstein-v4-Qwen-1.5-32B
 
 
 
 
 
 
 
130
 
131
- This model is a fine-tuned version of [Qwen/Qwen1.5-32B](https://huggingface.co/Qwen/Qwen1.5-32B) on the None dataset.
132
 
133
- ## Model description
134
 
135
- More information needed
 
136
 
137
- ## Intended uses & limitations
 
 
 
 
 
 
 
 
138
 
139
- More information needed
140
 
141
- ## Training and evaluation data
142
 
143
- More information needed
144
 
145
- ## Training procedure
146
 
147
- ### Training hyperparameters
148
 
149
- The following hyperparameters were used during training:
150
- - learning_rate: 0.0002
151
- - train_batch_size: 1
152
- - eval_batch_size: 1
153
- - seed: 42
154
- - distributed_type: multi-GPU
155
- - num_devices: 9
156
- - gradient_accumulation_steps: 4
157
- - total_train_batch_size: 36
158
- - total_eval_batch_size: 9
159
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
160
- - lr_scheduler_type: cosine
161
- - lr_scheduler_warmup_steps: 10
162
- - num_epochs: 2
163
 
164
- ### Training results
165
 
 
166
 
 
 
 
 
 
 
 
167
 
168
- ### Framework versions
169
 
170
- - PEFT 0.10.0
171
- - Transformers 4.40.0.dev0
172
- - Pytorch 2.1.2+cu118
173
- - Datasets 2.18.0
174
- - Tokenizers 0.15.0
 
1
  ---
2
+ language:
3
+ - en
4
  license: other
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
8
+ - phi
9
+ - phi2
10
+ - einstein
11
+ - instruct
12
+ - finetune
13
+ - chatml
14
+ - gpt4
15
+ - synthetic data
16
+ - science
17
+ - physics
18
+ - chemistry
19
+ - biology
20
+ - math
21
  base_model: Qwen/Qwen1.5-32B
22
+ datasets:
23
+ - allenai/ai2_arc
24
+ - camel-ai/physics
25
+ - camel-ai/chemistry
26
+ - camel-ai/biology
27
+ - camel-ai/math
28
+ - metaeval/reclor
29
+ - openbookqa
30
+ - mandyyyyii/scibench
31
+ - derek-thomas/ScienceQA
32
+ - TIGER-Lab/ScienceEval
33
+ - jondurbin/airoboros-3.2
34
+ - LDJnr/Capybara
35
+ - Cot-Alpaca-GPT4-From-OpenHermes-2.5
36
+ - STEM-AI-mtl/Electrical-engineering
37
+ - knowrohit07/saraswati-stem
38
+ - sablo/oasst2_curated
39
+ - glaiveai/glaive-code-assistant
40
+ - lmsys/lmsys-chat-1m
41
+ - TIGER-Lab/MathInstruct
42
+ - bigbio/med_qa
43
+ - meta-math/MetaMathQA-40K
44
+ - openbookqa
45
+ - piqa
46
+ - metaeval/reclor
47
+ - derek-thomas/ScienceQA
48
+ - scibench
49
+ - sciq
50
+ - Open-Orca/SlimOrca
51
+ - migtissera/Synthia-v1.3
52
+ - TIGER-Lab/ScienceEval
53
  ---
54
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/VTacthtA6N97SqD23WtwB.png)
55
 
56
+ # 🔬 Einstein-v4-Qwen-1.5-32B
57
+
58
+ This model is a [QLoRA](https://arxiv.org/abs/2305.14314) fine-tuned version of [Qwen/Qwen1.5-32B](https://huggingface.co/Qwen/Qwen1.5-32B) on diverse datasets.
59
+
60
+ This model is finetuned using `8xRTX3090` + `1xRTXA6000` using [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
61
+
62
+ This model's training was sponsored by [sablo.ai](https://sablo.ai).
63
 
 
64
  <details><summary>See axolotl config</summary>
65
 
66
  axolotl version: `0.4.0`
 
172
  unk_token: "<unk>"
173
  tokens:
174
  - "<|im_start|>"
175
+ ```
176
+ </details><br>
177
+
178
+ # 💬 Prompt Template
179
+
180
+ You can use this prompt template while using the model:
181
+
182
+ ### ChatML
183
 
184
+ ```
185
+ <|im_start|>system
186
+ {system}<|im_end|>
187
+ <|im_start|>user
188
+ {user}<|im_end|>
189
+ <|im_start|>assistant
190
+ {asistant}<|im_end|>
191
  ```
192
 
193
+ This prompt template is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating), which means you can format messages using the
194
+ `tokenizer.apply_chat_template()` method:
195
 
196
+ ```python
197
+ messages = [
198
+ {"role": "system", "content": "You are helpful AI asistant."},
199
+ {"role": "user", "content": "Hello!"}
200
+ ]
201
+ gen_input = tokenizer.apply_chat_template(message, return_tensors="pt")
202
+ model.generate(**gen_input)
203
+ ```
204
 
205
+ # 🔄 Quantizationed versions
206
 
207
+ Quantizationed versions of this model is currently not available.
208
 
209
+ # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
210
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v4-Qwen-1.5-32B)
211
 
212
+ | Metric |Value|
213
+ |---------------------------------|----:|
214
+ |Avg. |68.54|
215
+ |AI2 Reasoning Challenge (25-Shot)|62.37|
216
+ |HellaSwag (10-Shot) |83.85|
217
+ |MMLU (5-Shot) |74.04|
218
+ |TruthfulQA (0-shot) |58.86|
219
+ |Winogrande (5-shot) |80.43|
220
+ |GSM8k (5-shot) |51.71|
221
 
 
222
 
223
+ # 🤖 Additional information about training
224
 
225
+ This model is full fine-tuned for 2 epochs.
226
 
227
+ Total number of steps was 3352.
228
 
229
+ <details><summary>Loss graph</summary>
230
 
231
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/0Vp8iDmXi4-XbQCiwQtNP.png)
232
+
233
+ </details><br>
 
 
 
 
 
 
 
 
 
 
 
234
 
235
+ # 🤝 Acknowledgments
236
 
237
+ Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
238
 
239
+ Thanks to all the dataset authors mentioned in the datasets section.
240
+
241
+ Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
242
+
243
+ Thanks to all open source AI community.
244
+
245
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
246
 
247
+ If you would like to support me:
248
 
249
+ [☕ Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)