Update README.md
Browse files
README.md
CHANGED
@@ -23,7 +23,7 @@ This model uses an 1.3 billion parameters model as base to be further fine-tuned
|
|
23 |
This model was trained on less than 25% of the dataset yet achieves competitive performance to current sota on open llm leaderboard.
|
24 |
|
25 |
### training
|
26 |
-
Training took ~5 P100 gpu hours. It was trained on 15,000 gpt-4 shuffled samples.
|
27 |
|
28 |
### prompt
|
29 |
```
|
|
|
23 |
This model was trained on less than 25% of the dataset yet achieves competitive performance to current sota on open llm leaderboard.
|
24 |
|
25 |
### training
|
26 |
+
Training took ~5 P100 gpu hours. It was trained on 15,000 gpt-4 shuffled samples. no-prompt was fine-tuned using lower learning rates ensuring it keeps as much general knowledge as possible.
|
27 |
|
28 |
### prompt
|
29 |
```
|