wepolyu commited on
Commit
b151a9d
1 Parent(s): a125062

Model save

Browse files
README.md ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/flan-t5-large
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ - f1
9
+ - recall
10
+ - precision
11
+ model-index:
12
+ - name: KGAQ-2
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # KGAQ-2
20
+
21
+ This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on an unknown dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 6.0725
24
+ - Rouge1: 46.2376
25
+ - Rouge2: 21.4997
26
+ - Rougel: 39.6036
27
+ - Rougelsum: 46.3269
28
+ - Gen Len: 4.2121
29
+ - F1: 0.3205
30
+ - Recall: 0.6757
31
+ - Precision: 0.2101
32
+
33
+ ## Model description
34
+
35
+ More information needed
36
+
37
+ ## Intended uses & limitations
38
+
39
+ More information needed
40
+
41
+ ## Training and evaluation data
42
+
43
+ More information needed
44
+
45
+ ## Training procedure
46
+
47
+ ### Training hyperparameters
48
+
49
+ The following hyperparameters were used during training:
50
+ - learning_rate: 0.001
51
+ - train_batch_size: 8
52
+ - eval_batch_size: 8
53
+ - seed: 42
54
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
55
+ - lr_scheduler_type: cosine_with_restarts
56
+ - num_epochs: 10
57
+
58
+ ### Training results
59
+
60
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | F1 | Recall | Precision |
61
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|:------:|:------:|:---------:|
62
+ | No log | 1.0 | 50 | 3.1330 | 38.9228 | 19.8713 | 34.8665 | 39.0223 | 3.6162 | 0.3373 | 0.5957 | 0.2353 |
63
+ | No log | 2.0 | 100 | 3.2460 | 42.6051 | 20.0714 | 38.2234 | 42.823 | 3.9697 | 0.3275 | 0.5385 | 0.2353 |
64
+ | No log | 3.0 | 150 | 3.4413 | 42.2575 | 19.3868 | 36.9508 | 42.1996 | 4.1313 | 0.3415 | 0.6222 | 0.2353 |
65
+ | 1.9251 | 4.0 | 200 | 3.6553 | 41.9902 | 19.8751 | 36.961 | 42.1914 | 3.9899 | 0.3522 | 0.7 | 0.2353 |
66
+ | 1.9251 | 5.0 | 250 | 3.9188 | 41.6177 | 19.8385 | 36.9836 | 41.7831 | 4.0404 | 0.3648 | 0.725 | 0.2437 |
67
+ | 1.9251 | 6.0 | 300 | 4.0309 | 40.2818 | 15.9608 | 35.0963 | 40.3224 | 4.3838 | 0.3522 | 0.7 | 0.2353 |
68
+ | 1.9251 | 7.0 | 350 | 4.4151 | 40.1585 | 14.4247 | 34.3216 | 40.2886 | 4.3131 | 0.1185 | 0.5 | 0.0672 |
69
+ | 0.6344 | 8.0 | 400 | 4.9239 | 42.9643 | 19.2829 | 36.6803 | 43.0145 | 4.4646 | 0.3097 | 0.6667 | 0.2017 |
70
+ | 0.6344 | 9.0 | 450 | 5.9057 | 45.7386 | 21.5407 | 39.3743 | 45.7904 | 4.5253 | 0.3205 | 0.6757 | 0.2101 |
71
+ | 0.6344 | 10.0 | 500 | 6.0725 | 46.2376 | 21.4997 | 39.6036 | 46.3269 | 4.2121 | 0.3205 | 0.6757 | 0.2101 |
72
+
73
+
74
+ ### Framework versions
75
+
76
+ - Transformers 4.43.3
77
+ - Pytorch 2.3.1+cu121
78
+ - Datasets 2.20.0
79
+ - Tokenizers 0.19.1
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.43.3"
6
+ }
logs/events.out.tfevents.1722598928.c337331a3a44.16200.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f4737b0931ba01ed5d8f46edb6935d462d697d34dfb1bcee636b8e2af5dccc6
3
- size 11687
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:544ea68f37594820a8d09473cd7a57dc7c6f165cd6e2bf495d9b737ac9689c50
3
+ size 12715
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa16a9ec52dddf84264d6a538b7bc56d9ccb3f9592b56569f8d4206d3f6beb92
3
  size 3132668808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddcb55938072f59d1cad479d376c421fd3782f94bfc1e9e82f1d604e24a11f1c
3
  size 3132668808