Update README.md
Browse files
README.md
CHANGED
@@ -20,12 +20,14 @@ Colab: [link](https://colab.research.google.com/drive/18GG-zwxyGAHoVXghlOQONkxb_
|
|
20 |
Training code: [link](https://github.com/IlyaGusev/rulm/tree/master/self_instruct)
|
21 |
|
22 |
Versions:
|
23 |
-
- v2, April 8, 2023: A model trained with outputs capped at 800 tokens instead of 400.
|
|
|
24 |
- Final loss: 0.995
|
25 |
- Time: 38 hours, 35 minutes
|
26 |
- Hardware: 1x RTX 3090Ti
|
27 |
- Wandb link: [link](https://wandb.ai/ilyagusev/rulm_self_instruct/runs/wy66bwim/overview?workspace=user-ilyagusev)
|
28 |
-
- v1, April 2, 2023: Original model trained with 400 max tokens in outputs. Both Russian and English are supported, trained on the ru_turbo_alpaca and alpaca-cleaned.
|
|
|
29 |
- Final loss: 1.074
|
30 |
- Time: 53 hours, 53 minutes
|
31 |
- Hardware: 1x RTX 3090
|
|
|
20 |
Training code: [link](https://github.com/IlyaGusev/rulm/tree/master/self_instruct)
|
21 |
|
22 |
Versions:
|
23 |
+
- v2, April 8, 2023: A model trained with outputs capped at 800 tokens instead of 400.
|
24 |
+
- `revision="3dff86a20f80ba203fec6e986a3422b83822ca70"`
|
25 |
- Final loss: 0.995
|
26 |
- Time: 38 hours, 35 minutes
|
27 |
- Hardware: 1x RTX 3090Ti
|
28 |
- Wandb link: [link](https://wandb.ai/ilyagusev/rulm_self_instruct/runs/wy66bwim/overview?workspace=user-ilyagusev)
|
29 |
+
- v1, April 2, 2023: Original model trained with 400 max tokens in outputs. Both Russian and English are supported, trained on the ru_turbo_alpaca and alpaca-cleaned.
|
30 |
+
- `revision="97209c36c9261eb9527371a341c70f90c39ee158"`
|
31 |
- Final loss: 1.074
|
32 |
- Time: 53 hours, 53 minutes
|
33 |
- Hardware: 1x RTX 3090
|