lapp0 commited on
Commit
f70aff9
1 Parent(s): 54c7f91

End of training

Browse files
README.md ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: gpt2
3
+ library_name: Distily
4
+ license: mit
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: distily_modelcard_try
9
+ results: []
10
+ ---
11
+
12
+ # distily_modelcard_try
13
+
14
+ This student model is distilled from the teacher model [gpt2](https://huggingface.co/gpt2) using the dataset (unspecified).
15
+
16
+ The [Distily](https://github.com/lapp0/distily) library was used for this distillation.
17
+
18
+ It achieves the following results on the evaluation set:
19
+ - eval_enwikippl: 38656.0
20
+ - eval_frwikippl: 218112.0
21
+ - eval_zhwikippl: 54001664.0
22
+ - eval_tinystoriesppl: 12160.0
23
+ - eval_loss: 6.4375
24
+ - eval_runtime: 0.0668
25
+ - eval_samples_per_second: 29.948
26
+ - eval_steps_per_second: 14.974
27
+
28
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
+ should probably proofread and complete it, then remove this comment.
30
+
31
+ ## Model description
32
+
33
+ More information needed
34
+
35
+ ## Intended uses & limitations
36
+
37
+ More information needed
38
+
39
+ ## Training and evaluation data
40
+
41
+ More information needed
42
+ -->
43
+
44
+ ## Training procedure
45
+
46
+ ### Training hyperparameters
47
+
48
+ The following hyperparameters were used during training:
49
+ - distillation_objective: DistillationObjective(logits_loss_component=LossComponent(label=logits, weight=1, loss_fn=kl, layer_mapper=None, projector=None), hs_loss_component=LossComponent(label=hs, weight=0, loss_fn=mse, layer_mapper=last, projector=None), attn_loss_component=LossComponent(label=attn, weight=0, loss_fn=mse, layer_mapper=layer-2, projector=None))
50
+ - train_embeddings: True
51
+ - learning_rate: 0.0001
52
+ - train_batch_size: 16
53
+ - eval_batch_size: 8
54
+ - seed: 42
55
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
56
+ - lr_scheduler_type: constant
57
+ - lr_scheduler_warmup_ratio: 0.2
58
+ - num_epochs: 1.0
59
+
60
+ ### Resource Usage
61
+ Peak GPU Memory: 15.4263 GB
62
+
63
+ ### Eval-Phase Metrics
64
+ | step | epoch | enwikippl | frwikippl | loss | runtime | samples_per_second | steps_per_second | tinystoriesppl | zhwikippl |
65
+ | --- | --- | --- | --- | --- | --- | --- | --- | --- | --- |
66
+ | **teacher eval** | | 43.75 | 61.75 | | | | | 11.8125 | 19.125 |
67
+ | 0 | 0 | 738734374912.0 | 47828755808256.0 | 20.375 | 0.128 | 15.619 | 7.81 | 2617245696.0 | 12232066859008.0 |
68
+ | 10 | 1.0 | 38656.0 | 218112.0 | 6.4375 | 0.0668 | 29.948 | 14.974 | 12160.0 | 54001664.0 |
69
+
70
+ ### Framework versions
71
+ - Distily 0.2.0
72
+ - Transformers 4.44.0
73
+ - Pytorch 2.3.0
74
+ - Datasets 2.21.0
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.44.0"
6
+ }
logs/per_device_train_batch_size=16/events.out.tfevents.1724248995.f383272e719b ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85235a35b5393c4397a5476fdf12f88209b0d16c9ba37da0adbd3bc750ce7cc1
3
+ size 302