Update README.md
Browse files
README.md
CHANGED
@@ -9,14 +9,31 @@ language:
|
|
9 |
|
10 |
Fine tune of [Smaug 72b v0.1](https://huggingface.co/abacusai/Smaug-72B-v0.1) using an alpaca data set I have handy. The data is of planning and reasoning, which I use to help allow a model to break down a set of asks into a logical plan. For some odd reason it bumps the mmlu and winogrande? I would have expected the ARC to go up over those two, but this is often more of an artform than a science at times. All thanks to [Albacus.AI](https://huggingface.co/abacusai) for sharing their work.
|
11 |
|
|
|
|
|
12 |
![img](./alpaca_dragon.png)
|
13 |
|
14 |
|
15 |
|
|
|
16 |
## How to Get Started with the Model
|
17 |
|
18 |
Use the code below to get started with the model.
|
19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
## Evaluation
|
22 |
|
@@ -82,7 +99,7 @@ Use the code below to get started with the model.
|
|
82 |
| hendrycksTest-us_foreign_policy | 94.00 |
|
83 |
| hendrycksTest-virology | 57.23 |
|
84 |
| hendrycksTest-world_religions | 89.47 |
|
85 |
-
| truthfulqa:mc |
|
86 |
| winogrande | 86.03 |
|
87 |
| gsm8k | 77.63 |
|
88 |
|
|
|
9 |
|
10 |
Fine tune of [Smaug 72b v0.1](https://huggingface.co/abacusai/Smaug-72B-v0.1) using an alpaca data set I have handy. The data is of planning and reasoning, which I use to help allow a model to break down a set of asks into a logical plan. For some odd reason it bumps the mmlu and winogrande? I would have expected the ARC to go up over those two, but this is often more of an artform than a science at times. All thanks to [Albacus.AI](https://huggingface.co/abacusai) for sharing their work.
|
11 |
|
12 |
+
I used the same dataset in training one of my owl series [Strix Rufipes 70B](https://huggingface.co/ibivibiv/strix-rufipes-70b), which has worked well for planning out development tasks and other technical work.
|
13 |
+
|
14 |
![img](./alpaca_dragon.png)
|
15 |
|
16 |
|
17 |
|
18 |
+
|
19 |
## How to Get Started with the Model
|
20 |
|
21 |
Use the code below to get started with the model.
|
22 |
|
23 |
+
```
|
24 |
+
# Load model directly
|
25 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
26 |
+
|
27 |
+
tokenizer = AutoTokenizer.from_pretrained("ibivibiv/alpaca-dragon-72b-v1")
|
28 |
+
model = AutoModelForCausalLM.from_pretrained("ibivibiv/alpaca-dragon-72b-v1")
|
29 |
+
|
30 |
+
inputs = tokenizer("### Instruction: Create a plan for developing the game of snake in python using pygame.\n### Response:\n", return_tensors="pt", return_attention_mask=False)
|
31 |
+
|
32 |
+
outputs = model.generate(**inputs, max_length=200)
|
33 |
+
text = tokenizer.batch_decode(outputs)[0]
|
34 |
+
print(text)
|
35 |
+
```
|
36 |
+
|
37 |
|
38 |
## Evaluation
|
39 |
|
|
|
99 |
| hendrycksTest-us_foreign_policy | 94.00 |
|
100 |
| hendrycksTest-virology | 57.23 |
|
101 |
| hendrycksTest-world_religions | 89.47 |
|
102 |
+
| truthfulqa:mc | 72.6 |
|
103 |
| winogrande | 86.03 |
|
104 |
| gsm8k | 77.63 |
|
105 |
|