IlyaGusev commited on
Commit
dfc2610
1 Parent(s): 0123035

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -20,12 +20,14 @@ Colab: [link](https://colab.research.google.com/drive/18GG-zwxyGAHoVXghlOQONkxb_
20
  Training code: [link](https://github.com/IlyaGusev/rulm/tree/master/self_instruct)
21
 
22
  Versions:
23
- - v2, April 8, 2023: A model trained with outputs capped at 800 tokens instead of 400. `revision="3dff86a20f80ba203fec6e986a3422b83822ca70"`
 
24
  - Final loss: 0.995
25
  - Time: 38 hours, 35 minutes
26
  - Hardware: 1x RTX 3090Ti
27
  - Wandb link: [link](https://wandb.ai/ilyagusev/rulm_self_instruct/runs/wy66bwim/overview?workspace=user-ilyagusev)
28
- - v1, April 2, 2023: Original model trained with 400 max tokens in outputs. Both Russian and English are supported, trained on the ru_turbo_alpaca and alpaca-cleaned. `revision="97209c36c9261eb9527371a341c70f90c39ee158"`
 
29
  - Final loss: 1.074
30
  - Time: 53 hours, 53 minutes
31
  - Hardware: 1x RTX 3090
 
20
  Training code: [link](https://github.com/IlyaGusev/rulm/tree/master/self_instruct)
21
 
22
  Versions:
23
+ - v2, April 8, 2023: A model trained with outputs capped at 800 tokens instead of 400.
24
+ - `revision="3dff86a20f80ba203fec6e986a3422b83822ca70"`
25
  - Final loss: 0.995
26
  - Time: 38 hours, 35 minutes
27
  - Hardware: 1x RTX 3090Ti
28
  - Wandb link: [link](https://wandb.ai/ilyagusev/rulm_self_instruct/runs/wy66bwim/overview?workspace=user-ilyagusev)
29
+ - v1, April 2, 2023: Original model trained with 400 max tokens in outputs. Both Russian and English are supported, trained on the ru_turbo_alpaca and alpaca-cleaned.
30
+ - `revision="97209c36c9261eb9527371a341c70f90c39ee158"`
31
  - Final loss: 1.074
32
  - Time: 53 hours, 53 minutes
33
  - Hardware: 1x RTX 3090