paulh27 commited on
Commit
c6337a6
1 Parent(s): 46f16c2

End of training

Browse files
README.md CHANGED
@@ -3,9 +3,23 @@ license: apache-2.0
3
  base_model: google-t5/t5-small
4
  tags:
5
  - generated_from_trainer
 
 
 
 
6
  model-index:
7
  - name: xsum_aligned_smallT5
8
- results: []
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,7 +27,14 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # xsum_aligned_smallT5
15
 
16
- This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on an unknown dataset.
 
 
 
 
 
 
 
17
 
18
  ## Model description
19
 
 
3
  base_model: google-t5/t5-small
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - lilferrit/xsum_t5_distillation
8
+ metrics:
9
+ - rouge
10
  model-index:
11
  - name: xsum_aligned_smallT5
12
+ results:
13
+ - task:
14
+ name: Summarization
15
+ type: summarization
16
+ dataset:
17
+ name: lilferrit/xsum_t5_distillation
18
+ type: lilferrit/xsum_t5_distillation
19
+ metrics:
20
+ - name: Rouge1
21
+ type: rouge
22
+ value: 28.6381
23
  ---
24
 
25
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
27
 
28
  # xsum_aligned_smallT5
29
 
30
+ This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on the lilferrit/xsum_t5_distillation dataset.
31
+ It achieves the following results on the evaluation set:
32
+ - Loss: 2.5258
33
+ - Rouge1: 28.6381
34
+ - Rouge2: 7.1512
35
+ - Rougel: 21.3477
36
+ - Rougelsum: 21.2928
37
+ - Gen Len: 27.92
38
 
39
  ## Model description
40
 
all_results.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 30.77,
3
+ "eval_gen_len": 27.92,
4
+ "eval_loss": 2.5258138179779053,
5
+ "eval_rouge1": 28.6381,
6
+ "eval_rouge2": 7.1512,
7
+ "eval_rougeL": 21.3477,
8
+ "eval_rougeLsum": 21.2928,
9
+ "eval_runtime": 34.7229,
10
+ "eval_samples": 100,
11
+ "eval_samples_per_second": 2.88,
12
+ "eval_steps_per_second": 2.88,
13
+ "train_loss": 1.359188232421875,
14
+ "train_runtime": 123.5443,
15
+ "train_samples": 100,
16
+ "train_samples_per_second": 25.902,
17
+ "train_steps_per_second": 1.619
18
+ }
eval_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 30.77,
3
+ "eval_gen_len": 27.92,
4
+ "eval_loss": 2.5258138179779053,
5
+ "eval_rouge1": 28.6381,
6
+ "eval_rouge2": 7.1512,
7
+ "eval_rougeL": 21.3477,
8
+ "eval_rougeLsum": 21.2928,
9
+ "eval_runtime": 34.7229,
10
+ "eval_samples": 100,
11
+ "eval_samples_per_second": 2.88,
12
+ "eval_steps_per_second": 2.88
13
+ }
generation_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_from_model_config": true,
3
  "decoder_start_token_id": 0,
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
 
1
  {
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
runs/Apr22_22-06-47_nlpg03.cs.washington.edu/events.out.tfevents.1713848986.nlpg03.cs.washington.edu.1675484.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fa79e977bdab6d9494f6c14768772155195d8389fdd4c282fb24510d6161e8b
3
+ size 613
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 30.77,
3
+ "train_loss": 1.359188232421875,
4
+ "train_runtime": 123.5443,
5
+ "train_samples": 100,
6
+ "train_samples_per_second": 25.902,
7
+ "train_steps_per_second": 1.619
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 30.76923076923077,
5
+ "eval_steps": 500,
6
+ "global_step": 200,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 15.38,
13
+ "grad_norm": 2.1158251762390137,
14
+ "learning_rate": 0.0001,
15
+ "loss": 1.8108,
16
+ "step": 100
17
+ },
18
+ {
19
+ "epoch": 30.77,
20
+ "grad_norm": 2.3235597610473633,
21
+ "learning_rate": 0.0,
22
+ "loss": 0.9076,
23
+ "step": 200
24
+ },
25
+ {
26
+ "epoch": 30.77,
27
+ "step": 200,
28
+ "total_flos": 752998431719424.0,
29
+ "train_loss": 1.359188232421875,
30
+ "train_runtime": 123.5443,
31
+ "train_samples_per_second": 25.902,
32
+ "train_steps_per_second": 1.619
33
+ }
34
+ ],
35
+ "logging_steps": 100,
36
+ "max_steps": 200,
37
+ "num_input_tokens_seen": 0,
38
+ "num_train_epochs": 34,
39
+ "save_steps": 500,
40
+ "total_flos": 752998431719424.0,
41
+ "train_batch_size": 8,
42
+ "trial_name": null,
43
+ "trial_params": null
44
+ }