alpindale commited on
Commit
678e4f4
1 Parent(s): b0e46e9

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: tongyi-qianwen
4
+ license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE
5
+ language:
6
+ - en
7
+ - zh
8
+ pipeline_tag: text-generation
9
+ tags:
10
+ - chat
11
+ ---
12
+
13
+ ![](https://files.catbox.moe/ngqnb1.png)
14
+
15
+ This is the first in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of [Qwen-2 72B Instruct](https://huggingface.co/Qwen/Qwen2-72B-Instruct).
16
+
17
+
18
+ ## Prompting
19
+ Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
20
+
21
+ ```py
22
+ """<|im_start|>user
23
+ Hi there!<|im_end|>
24
+ <|im_start|>assistant
25
+ Nice to meet you!<|im_end|>
26
+ <|im_start|>user
27
+ Can I ask a question?<|im_end|>
28
+ <|im_start|>assistant
29
+ """
30
+ ```
31
+
32
+ ## Credits
33
+
34
+ This model has been a team effort, credits go to:
35
+
36
+ - [Sao10K](https://huggingface.co/Sao10K) for help with (and cleaning up!) the dataset.
37
+ - [alpindale](https://huggingface.co/alpindale) for the training.
38
+ - [kalomaze](https://huggingface.co/kalomaze) for helping with the hyperparameter tuning.
39
+ - Various other people for their continued help as we tuned the parameters, restarted failed runs. In no particular order: [Doctor Shotgun](https://huggingface.co/Doctor-Shotgun), [Lucy](https://huggingface.co/lucyknada), [Nopm](https://huggingface.co/nopm), [Mango](https://huggingface.co/MangoMango69420), and the rest of the Silly Tilly.
40
+
41
+ And last but not least, we'd like to thank [Kearm](https://twitter.com/Nottlespike) for sponsoring the compute needed to train this model.
42
+
43
+ ## Training
44
+ The training was done with 55 million tokens of high-quality RP data, over 1.5 epochs. We used 8x [AMD Instinct™ MI300X Accelerators](https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html) for the full-parameter fine-tuning of the model.
45
+
46
+
47
+ ## Safety
48
+ ...