lapp0 commited on
Commit
113b768
1 Parent(s): b378d0c

End of training

Browse files
README.md CHANGED
@@ -16,14 +16,14 @@ This student model is distilled from the teacher model [gpt2](https://huggingfac
16
  The [Distily](https://github.com/lapp0/distily) library was used for this distillation.
17
 
18
  It achieves the following results on the evaluation set:
19
- - eval_enwikippl: 2192.0
20
- - eval_frwikippl: 11200.0
21
- - eval_zhwikippl: 93184.0
22
- - eval_tinystoriesppl: 1808.0
23
- - eval_loss: 2.6293
24
- - eval_runtime: 16.9228
25
- - eval_samples_per_second: 59.092
26
- - eval_steps_per_second: 7.386
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment.
@@ -64,20 +64,20 @@ Peak GPU Memory: 7.9368 GB
64
  | step | epoch | enwikippl | frwikippl | loss | runtime | samples_per_second | steps_per_second | tinystoriesppl | zhwikippl |
65
  | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
66
  | **teacher eval** | | 43.75 | 61.75 | | | | | 11.8125 | 19.125 |
67
- | 0 | 0 | 2473901162496.0 | 170424302305280.0 | 20.7680 | 16.794 | 59.545 | 7.443 | 4060086272.0 | 71468255805440.0 |
68
- | 1000 | 0.0808 | 688.0 | 3728.0 | 1.9530 | 16.821 | 59.449 | 7.431 | 652.0 | 2784.0 |
69
- | 2000 | 0.1616 | 1728.0 | 8256.0 | 2.4948 | 16.7878 | 59.567 | 7.446 | 1384.0 | 35584.0 |
70
- | 3000 | 0.2424 | 2040.0 | 10112.0 | 2.6087 | 16.7522 | 59.694 | 7.462 | 1720.0 | 64256.0 |
71
- | 4000 | 0.3232 | 2160.0 | 9280.0 | 2.6353 | 16.796 | 59.538 | 7.442 | 1816.0 | 57088.0 |
72
- | 5000 | 0.4040 | 1904.0 | 9088.0 | 2.5782 | 16.8206 | 59.451 | 7.431 | 1848.0 | 61440.0 |
73
- | 6000 | 0.4848 | 1840.0 | 8960.0 | 2.5344 | 16.7618 | 59.659 | 7.457 | 1592.0 | 69120.0 |
74
- | 7000 | 0.5657 | 1808.0 | 8512.0 | 2.5269 | 16.7913 | 59.555 | 7.444 | 1648.0 | 60672.0 |
75
- | 8000 | 0.6465 | 2096.0 | 8960.0 | 2.6404 | 16.8233 | 59.442 | 7.43 | 1928.0 | 137216.0 |
76
- | 9000 | 0.7273 | 2192.0 | 11200.0 | 2.6293 | 16.9228 | 59.092 | 7.386 | 1808.0 | 93184.0 |
77
- | 10000 | 0.8081 | 1944.0 | 9984.0 | 2.5759 | 16.857 | 59.323 | 7.415 | 1568.0 | 80896.0 |
78
- | 11000 | 0.8889 | 1736.0 | 9344.0 | 2.5147 | 16.8438 | 59.369 | 7.421 | 1488.0 | 48640.0 |
79
- | 12000 | 0.9697 | 2224.0 | 11840.0 | 2.6633 | 16.7839 | 59.581 | 7.448 | 1968.0 | 98816.0 |
80
- | 12375 | 1.0 | 2432.0 | 11072.0 | 2.7197 | 16.7952 | 59.541 | 7.443 | 2176.0 | 109568.0 |
81
 
82
  ### Framework versions
83
  - Distily 0.2.0
 
16
  The [Distily](https://github.com/lapp0/distily) library was used for this distillation.
17
 
18
  It achieves the following results on the evaluation set:
19
+ - eval_enwikippl: 5056.0
20
+ - eval_frwikippl: 3696.0
21
+ - eval_zhwikippl: 29312.0
22
+ - eval_tinystoriesppl: 4672.0
23
+ - eval_loss: 1.3042
24
+ - eval_runtime: 16.773
25
+ - eval_samples_per_second: 59.62
26
+ - eval_steps_per_second: 7.452
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
  should probably proofread and complete it, then remove this comment.
 
64
  | step | epoch | enwikippl | frwikippl | loss | runtime | samples_per_second | steps_per_second | tinystoriesppl | zhwikippl |
65
  | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
66
  | **teacher eval** | | 43.75 | 61.75 | | | | | 11.8125 | 19.125 |
67
+ | 0 | 0 | 2336462209024.0 | 122045790683136.0 | 24.1200 | 16.7622 | 59.658 | 7.457 | 4429185024.0 | 25975962206208.0 |
68
+ | 1000 | 0.0808 | 3312.0 | 3184.0 | 1.1229 | 16.7657 | 59.646 | 7.456 | 2720.0 | 2992.0 |
69
+ | 2000 | 0.1616 | 5248.0 | 4048.0 | 1.2528 | 16.7825 | 59.586 | 7.448 | 5408.0 | 9984.0 |
70
+ | 3000 | 0.2424 | 5600.0 | 3744.0 | 1.2812 | 16.7695 | 59.632 | 7.454 | 5312.0 | 23680.0 |
71
+ | 4000 | 0.3232 | 5408.0 | 3920.0 | 1.2832 | 16.8694 | 59.279 | 7.41 | 5440.0 | 33280.0 |
72
+ | 5000 | 0.4040 | 5376.0 | 3952.0 | 1.2841 | 16.7361 | 59.751 | 7.469 | 5408.0 | 27008.0 |
73
+ | 6000 | 0.4848 | 5344.0 | 3680.0 | 1.2770 | 16.7635 | 59.653 | 7.457 | 5440.0 | 29312.0 |
74
+ | 7000 | 0.5657 | 5024.0 | 3760.0 | 1.2800 | 16.7492 | 59.704 | 7.463 | 5184.0 | 39936.0 |
75
+ | 8000 | 0.6465 | 4992.0 | 3712.0 | 1.2922 | 16.7445 | 59.721 | 7.465 | 5088.0 | 26752.0 |
76
+ | 9000 | 0.7273 | 5056.0 | 3696.0 | 1.3042 | 16.773 | 59.62 | 7.452 | 4672.0 | 29312.0 |
77
+ | 10000 | 0.8081 | 5824.0 | 3648.0 | 1.3192 | 16.7669 | 59.641 | 7.455 | 5312.0 | 24448.0 |
78
+ | 11000 | 0.8889 | 5568.0 | 3872.0 | 1.3215 | 16.8215 | 59.448 | 7.431 | 5504.0 | 40704.0 |
79
+ | 12000 | 0.9697 | 5440.0 | 3792.0 | 1.3263 | 16.7825 | 59.586 | 7.448 | 5120.0 | 72704.0 |
80
+ | 12375 | 1.0 | 5696.0 | 3936.0 | 1.3389 | 16.7852 | 59.576 | 7.447 | 5696.0 | 40192.0 |
81
 
82
  ### Framework versions
83
  - Distily 0.2.0
logs/completed.flag ADDED
File without changes
logs/dataset_subset=default, dataset_uri=distily_c4_multilingual_1M/events.out.tfevents.1724131499.02dbb11e2dcc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec817be5a82bd8f62a468d4f1e8d34c55c85f6516690074f226ff4730f676e33
3
+ size 5852968
logs/dataset_subset=default, dataset_uri=distily_c4_multilingual_1M/events.out.tfevents.1724135677.02dbb11e2dcc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:063d08fdd0959113009ef7212f30778cba39b229d1b355b4fa1f5eb7a8820a88
3
+ size 307
logs/events.out.tfevents.1724131158.02dbb11e2dcc CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12278b9d145fd2c6847dff7dea295578db7bd43aee4be002f23b4f56cd9ce1a0
3
- size 307
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:323bccb39f408e384a63c920f2c1fffe8a29e3beb175b2e3ba021879b695c7cd
3
+ size 578
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:776f3531eec7b3712662c7d587fe16cf37bc93e8816939f74bf1498055406a03
3
  size 248894656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5720c9d9de3ebbbaf9349e73e056d8dbd0e660e232ace49a07ac1d51bef10b1
3
  size 248894656
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7777a3b236d5a2940cd4ae7de66e1e80e17576a70be7777d54114b4ecf4ff248
3
- size 1017899080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:430090f16bc1caaa707b1032e32819ab7c66f4a1d6225492b346dde37cd48308
3
+ size 1017899144