|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- fblgit/simple-math |
|
- jondurbin/bagel-v0.3 |
|
base_model: abacusai/Smaug-34B-v0.1 |
|
tags: |
|
- UNA |
|
- simple-math |
|
- juanako |
|
--- |
|
|
|
# UNA-SimpleSmaug-34b-v1beta |
|
|
|
Scoring 04-February-2024 #1 34B model, outperforming its original base model Smaug-34B-v0.1 with `77.41` ๐ |
|
Oh, btw.. this one went thru SFT so the abacus inside Smaug is back to normal.. so you can further train/dpo him .. RESET! |
|
|
|
![UNA](https://huggingface.co/fblgit/UNA-SimpleSmaug-34b-v1beta/resolve/main/unasimple.png) |
|
Applied UNA only on the Attention, not on the MLP's |
|
* Is based on Smaug |
|
* SimpleMath dataset |
|
* It was trained on Axolotl |
|
|
|
## Experiment |
|
The thing here is to understand whats the impact of SimpleMath applied at the attention layer during a SFT session and how it impacts on the neural network overall. |
|
|
|
Results: Improving mathematican and reasoning capabilities without degrading and presserving previous training sessions. |
|
|
|
## Evals |
|
|
|
Pending, but so far this one |
|
``` |
|
| Task |Version| Metric |Value | |
|
|-------------|------:|--------|----------------:| |
|
|arc_challenge| HF|acc_norm| 0.7457337883959 | |
|
|gsm8k | HF|acc | 0.7247915087187 | |
|
|mmlu | HF|acc | 0.7649553475572 | |
|
|mmlu | HF|acc_norm| 0.7681713551647 | |
|
|hellaswag | HF|acc_norm| 0.8673571001792 | |
|
|truthfulqa | HF|mc2 | 0.7016557407771 | |
|
|winogrande | HF|acc | 0.8382004735595 | |
|
|------------------------------------------------| |
|
``` |
|
|
|
Increasing GSM, MMLU, ARC, WINO. |
|
|
|
## Citations |
|
To abacusai for making Smaug-34B, the Bagel, and all the magic behind the base model. |
|
|
|
If you use the model, provide citation even for merges or anything. |
|
And enjoy our ModelSimilarities tool detector https://github.com/fblgit/model-similarity |