juny76 commited on
Commit
078a032
·
verified ·
1 Parent(s): 99ba377

Upload 11 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
gliner_config.json ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "class_token_index": -1,
3
+ "dropout": 0.3,
4
+ "embed_ent_token": true,
5
+ "encoder_config": {
6
+ "_name_or_path": "microsoft/deberta-v3-base",
7
+ "add_cross_attention": false,
8
+ "architectures": null,
9
+ "attention_probs_dropout_prob": 0.1,
10
+ "bad_words_ids": null,
11
+ "begin_suppress_tokens": null,
12
+ "bos_token_id": null,
13
+ "chunk_size_feed_forward": 0,
14
+ "cross_attention_hidden_size": null,
15
+ "decoder_start_token_id": null,
16
+ "diversity_penalty": 0,
17
+ "do_sample": false,
18
+ "early_stopping": false,
19
+ "encoder_no_repeat_ngram_size": 0,
20
+ "eos_token_id": null,
21
+ "exponential_decay_length_penalty": null,
22
+ "finetuning_task": null,
23
+ "forced_bos_token_id": null,
24
+ "forced_eos_token_id": null,
25
+ "hidden_act": "gelu",
26
+ "hidden_dropout_prob": 0.1,
27
+ "hidden_size": 768,
28
+ "id2label": {
29
+ "0": "LABEL_0",
30
+ "1": "LABEL_1"
31
+ },
32
+ "initializer_range": 0.02,
33
+ "intermediate_size": 3072,
34
+ "is_decoder": false,
35
+ "is_encoder_decoder": false,
36
+ "label2id": {
37
+ "LABEL_0": 0,
38
+ "LABEL_1": 1
39
+ },
40
+ "layer_norm_eps": 1e-7,
41
+ "length_penalty": 1,
42
+ "max_length": 20,
43
+ "max_position_embeddings": 512,
44
+ "max_relative_positions": -1,
45
+ "min_length": 0,
46
+ "model_type": "deberta-v2",
47
+ "no_repeat_ngram_size": 0,
48
+ "norm_rel_ebd": "layer_norm",
49
+ "num_attention_heads": 12,
50
+ "num_beam_groups": 1,
51
+ "num_beams": 1,
52
+ "num_hidden_layers": 12,
53
+ "num_return_sequences": 1,
54
+ "output_attentions": false,
55
+ "output_hidden_states": false,
56
+ "output_scores": false,
57
+ "pad_token_id": 0,
58
+ "pooler_dropout": 0,
59
+ "pooler_hidden_act": "gelu",
60
+ "pooler_hidden_size": 768,
61
+ "pos_att_type": [
62
+ "p2c",
63
+ "c2p"
64
+ ],
65
+ "position_biased_input": false,
66
+ "position_buckets": 256,
67
+ "prefix": null,
68
+ "problem_type": null,
69
+ "pruned_heads": {},
70
+ "relative_attention": true,
71
+ "remove_invalid_values": false,
72
+ "repetition_penalty": 1,
73
+ "return_dict": true,
74
+ "return_dict_in_generate": false,
75
+ "sep_token_id": null,
76
+ "share_att_key": true,
77
+ "suppress_tokens": null,
78
+ "task_specific_params": null,
79
+ "temperature": 1,
80
+ "tf_legacy_loss": false,
81
+ "tie_encoder_decoder": false,
82
+ "tie_word_embeddings": true,
83
+ "tokenizer_class": null,
84
+ "top_k": 50,
85
+ "top_p": 1,
86
+ "torch_dtype": null,
87
+ "torchscript": false,
88
+ "type_vocab_size": 0,
89
+ "typical_p": 1,
90
+ "use_bfloat16": false,
91
+ "vocab_size": 128100
92
+ },
93
+ "ent_token": "<<ENT>>",
94
+ "eval_every": 5000,
95
+ "fine_tune": true,
96
+ "freeze_token_rep": false,
97
+ "fuse_layers": false,
98
+ "has_rnn": true,
99
+ "hidden_size": 768,
100
+ "label_smoothing": 0,
101
+ "labels_encoder": "BAAI/bge-small-en-v1.5",
102
+ "labels_encoder_config": {
103
+ "_name_or_path": "BAAI/bge-small-en-v1.5",
104
+ "add_cross_attention": false,
105
+ "architectures": [
106
+ "BertModel"
107
+ ],
108
+ "attention_probs_dropout_prob": 0.1,
109
+ "bad_words_ids": null,
110
+ "begin_suppress_tokens": null,
111
+ "bos_token_id": null,
112
+ "chunk_size_feed_forward": 0,
113
+ "classifier_dropout": null,
114
+ "cross_attention_hidden_size": null,
115
+ "decoder_start_token_id": null,
116
+ "diversity_penalty": 0,
117
+ "do_sample": false,
118
+ "early_stopping": false,
119
+ "encoder_no_repeat_ngram_size": 0,
120
+ "eos_token_id": null,
121
+ "exponential_decay_length_penalty": null,
122
+ "finetuning_task": null,
123
+ "forced_bos_token_id": null,
124
+ "forced_eos_token_id": null,
125
+ "hidden_act": "gelu",
126
+ "hidden_dropout_prob": 0.1,
127
+ "hidden_size": 384,
128
+ "id2label": {
129
+ "0": "LABEL_0"
130
+ },
131
+ "initializer_range": 0.02,
132
+ "intermediate_size": 1536,
133
+ "is_decoder": false,
134
+ "is_encoder_decoder": false,
135
+ "label2id": {
136
+ "LABEL_0": 0
137
+ },
138
+ "layer_norm_eps": 1e-12,
139
+ "length_penalty": 1,
140
+ "max_length": 20,
141
+ "max_position_embeddings": 512,
142
+ "min_length": 0,
143
+ "model_type": "bert",
144
+ "no_repeat_ngram_size": 0,
145
+ "num_attention_heads": 12,
146
+ "num_beam_groups": 1,
147
+ "num_beams": 1,
148
+ "num_hidden_layers": 12,
149
+ "num_return_sequences": 1,
150
+ "output_attentions": false,
151
+ "output_hidden_states": false,
152
+ "output_scores": false,
153
+ "pad_token_id": 0,
154
+ "position_embedding_type": "absolute",
155
+ "prefix": null,
156
+ "problem_type": null,
157
+ "pruned_heads": {},
158
+ "remove_invalid_values": false,
159
+ "repetition_penalty": 1,
160
+ "return_dict": true,
161
+ "return_dict_in_generate": false,
162
+ "sep_token_id": null,
163
+ "suppress_tokens": null,
164
+ "task_specific_params": null,
165
+ "temperature": 1,
166
+ "tf_legacy_loss": false,
167
+ "tie_encoder_decoder": false,
168
+ "tie_word_embeddings": true,
169
+ "tokenizer_class": null,
170
+ "top_k": 50,
171
+ "top_p": 1,
172
+ "torch_dtype": "float32",
173
+ "torchscript": false,
174
+ "type_vocab_size": 2,
175
+ "typical_p": 1,
176
+ "use_bfloat16": false,
177
+ "use_cache": true,
178
+ "vocab_size": 30522
179
+ },
180
+ "log_dir": "deberta/",
181
+ "loss_alpha": 0.8,
182
+ "loss_gamma": 2,
183
+ "loss_reduction": "sum",
184
+ "lr_encoder": "1e-5",
185
+ "lr_others": "5e-5",
186
+ "max_grad_norm": 10,
187
+ "max_len": 512,
188
+ "max_neg_type_ratio": 1,
189
+ "max_types": 100,
190
+ "max_width": 12,
191
+ "model_name": "microsoft/deberta-v3-base",
192
+ "model_type": "gliner",
193
+ "name": "span level gliner",
194
+ "num_post_fusion_layers": 1,
195
+ "num_steps": 100000,
196
+ "post_fusion_schema": "",
197
+ "prev_path": null,
198
+ "random_drop": true,
199
+ "root_dir": "gliner_logs",
200
+ "save_total_limit": 3,
201
+ "scheduler_type": "cosine",
202
+ "sep_token": "<<SEP>>",
203
+ "shuffle_types": true,
204
+ "size_sup": -1,
205
+ "span_mode": "markerV0",
206
+ "subtoken_pooling": "first",
207
+ "train_batch_size": 8,
208
+ "train_data": "data/nuner_train.json",
209
+ "transformers_version": "4.45.2",
210
+ "val_data_dir": "none",
211
+ "vocab_size": -1,
212
+ "warmup_ratio": 0.05,
213
+ "weight_decay_encoder": 0.1,
214
+ "weight_decay_other": 0.1,
215
+ "words_splitter_type": "whitespace"
216
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed884e4c7dfd04f844f34450ae3dea9fa4dde54141913f2e58b8237988643aa0
3
+ size 1909053417
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b05009068f723953cd2a19492413f0842554ace11fa073b7996c6ded005136f
3
+ size 969281034
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7b36b09dc0d742cdacb85dd59355e8b0ba4e300d7b51f6f0b3f8c4b0262325d
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f1828686b8230c9d213a7809ad8e1222b72b42341bfd713e543489ff5454fd5
3
+ size 1064
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:335e95775c824ea2e442870aa5454366947f1d99fa433870300fec9cd5b62846
3
+ size 12745761
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "model_max_length": 1e+30,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
+ }
trainer_state.json ADDED
@@ -0,0 +1,303 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.568065810991625,
5
+ "eval_steps": 500,
6
+ "global_step": 9000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.19822587838842362,
13
+ "grad_norm": 47.1356201171875,
14
+ "learning_rate": 0.0000024777006937561945,
15
+ "loss": 21.4029,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.19822587838842362,
20
+ "eval_loss": 586.7496948242188,
21
+ "eval_runtime": 573.4198,
22
+ "eval_samples_per_second": 53,
23
+ "eval_steps_per_second": 6.625,
24
+ "step": 500
25
+ },
26
+ {
27
+ "epoch": 0.39645175677684724,
28
+ "grad_norm": 50.89448165893555,
29
+ "learning_rate": 0.000004955401387512389,
30
+ "loss": 11.7925,
31
+ "step": 1000
32
+ },
33
+ {
34
+ "epoch": 0.39645175677684724,
35
+ "eval_loss": 584.6571044921875,
36
+ "eval_runtime": 571.5589,
37
+ "eval_samples_per_second": 53.172,
38
+ "eval_steps_per_second": 6.647,
39
+ "step": 1000
40
+ },
41
+ {
42
+ "epoch": 0.5946776351652708,
43
+ "grad_norm": 56.098140716552734,
44
+ "learning_rate": 0.000007433102081268584,
45
+ "loss": 9.7243,
46
+ "step": 1500
47
+ },
48
+ {
49
+ "epoch": 0.5946776351652708,
50
+ "eval_loss": 438.975830078125,
51
+ "eval_runtime": 571.006,
52
+ "eval_samples_per_second": 53.224,
53
+ "eval_steps_per_second": 6.653,
54
+ "step": 1500
55
+ },
56
+ {
57
+ "epoch": 0.7929035135536945,
58
+ "grad_norm": 49.88700485229492,
59
+ "learning_rate": 0.000009910802775024778,
60
+ "loss": 8.6629,
61
+ "step": 2000
62
+ },
63
+ {
64
+ "epoch": 0.7929035135536945,
65
+ "eval_loss": 443.8860778808594,
66
+ "eval_runtime": 569.8536,
67
+ "eval_samples_per_second": 53.331,
68
+ "eval_steps_per_second": 6.667,
69
+ "step": 2000
70
+ },
71
+ {
72
+ "epoch": 0.991129391942118,
73
+ "grad_norm": 37.3896484375,
74
+ "learning_rate": 0.00000991223686038578,
75
+ "loss": 7.7527,
76
+ "step": 2500
77
+ },
78
+ {
79
+ "epoch": 0.991129391942118,
80
+ "eval_loss": 359.1586608886719,
81
+ "eval_runtime": 570.3107,
82
+ "eval_samples_per_second": 53.289,
83
+ "eval_steps_per_second": 6.661,
84
+ "step": 2500
85
+ },
86
+ {
87
+ "epoch": 1.1893552703305417,
88
+ "grad_norm": 45.05537414550781,
89
+ "learning_rate": 0.000009639072360499928,
90
+ "loss": 7.0532,
91
+ "step": 3000
92
+ },
93
+ {
94
+ "epoch": 1.1893552703305417,
95
+ "eval_loss": 297.4908447265625,
96
+ "eval_runtime": 573.7765,
97
+ "eval_samples_per_second": 52.967,
98
+ "eval_steps_per_second": 6.621,
99
+ "step": 3000
100
+ },
101
+ {
102
+ "epoch": 1.3875811487189653,
103
+ "grad_norm": 42.01622772216797,
104
+ "learning_rate": 0.000009190700553732768,
105
+ "loss": 6.8798,
106
+ "step": 3500
107
+ },
108
+ {
109
+ "epoch": 1.3875811487189653,
110
+ "eval_loss": 272.1216125488281,
111
+ "eval_runtime": 570.5906,
112
+ "eval_samples_per_second": 53.262,
113
+ "eval_steps_per_second": 6.658,
114
+ "step": 3500
115
+ },
116
+ {
117
+ "epoch": 1.5858070271073887,
118
+ "grad_norm": 26.336191177368164,
119
+ "learning_rate": 0.000008584055432624292,
120
+ "loss": 6.5736,
121
+ "step": 4000
122
+ },
123
+ {
124
+ "epoch": 1.5858070271073887,
125
+ "eval_loss": 279.0899353027344,
126
+ "eval_runtime": 578.6986,
127
+ "eval_samples_per_second": 52.516,
128
+ "eval_steps_per_second": 6.565,
129
+ "step": 4000
130
+ },
131
+ {
132
+ "epoch": 1.7840329054958124,
133
+ "grad_norm": 35.772953033447266,
134
+ "learning_rate": 0.00000784204861605915,
135
+ "loss": 6.3985,
136
+ "step": 4500
137
+ },
138
+ {
139
+ "epoch": 1.7840329054958124,
140
+ "eval_loss": 221.53627014160156,
141
+ "eval_runtime": 569.7883,
142
+ "eval_samples_per_second": 53.337,
143
+ "eval_steps_per_second": 6.667,
144
+ "step": 4500
145
+ },
146
+ {
147
+ "epoch": 1.982258783884236,
148
+ "grad_norm": 34.2270622253418,
149
+ "learning_rate": 0.000006992704029062722,
150
+ "loss": 6.0866,
151
+ "step": 5000
152
+ },
153
+ {
154
+ "epoch": 1.982258783884236,
155
+ "eval_loss": 190.98533630371094,
156
+ "eval_runtime": 571.0123,
157
+ "eval_samples_per_second": 53.223,
158
+ "eval_steps_per_second": 6.653,
159
+ "step": 5000
160
+ },
161
+ {
162
+ "epoch": 2.1804846622726597,
163
+ "grad_norm": 34.992889404296875,
164
+ "learning_rate": 0.000006068099502317051,
165
+ "loss": 5.732,
166
+ "step": 5500
167
+ },
168
+ {
169
+ "epoch": 2.1804846622726597,
170
+ "eval_loss": 225.56539916992188,
171
+ "eval_runtime": 570.6646,
172
+ "eval_samples_per_second": 53.255,
173
+ "eval_steps_per_second": 6.657,
174
+ "step": 5500
175
+ },
176
+ {
177
+ "epoch": 2.3787105406610833,
178
+ "grad_norm": 35.05290222167969,
179
+ "learning_rate": 0.000005103155264797376,
180
+ "loss": 5.8384,
181
+ "step": 6000
182
+ },
183
+ {
184
+ "epoch": 2.3787105406610833,
185
+ "eval_loss": 206.36277770996094,
186
+ "eval_runtime": 569.1115,
187
+ "eval_samples_per_second": 53.401,
188
+ "eval_steps_per_second": 6.675,
189
+ "step": 6000
190
+ },
191
+ {
192
+ "epoch": 2.576936419049507,
193
+ "grad_norm": 38.268341064453125,
194
+ "learning_rate": 0.0000041343150854317275,
195
+ "loss": 5.6189,
196
+ "step": 6500
197
+ },
198
+ {
199
+ "epoch": 2.576936419049507,
200
+ "eval_loss": 211.91796875,
201
+ "eval_runtime": 569.7007,
202
+ "eval_samples_per_second": 53.346,
203
+ "eval_steps_per_second": 6.668,
204
+ "step": 6500
205
+ },
206
+ {
207
+ "epoch": 2.7751622974379306,
208
+ "grad_norm": 42.83818054199219,
209
+ "learning_rate": 0.0000031981698740904465,
210
+ "loss": 5.5445,
211
+ "step": 7000
212
+ },
213
+ {
214
+ "epoch": 2.7751622974379306,
215
+ "eval_loss": 199.4452362060547,
216
+ "eval_runtime": 570.9073,
217
+ "eval_samples_per_second": 53.233,
218
+ "eval_steps_per_second": 6.654,
219
+ "step": 7000
220
+ },
221
+ {
222
+ "epoch": 2.9733881758263543,
223
+ "grad_norm": 42.24338150024414,
224
+ "learning_rate": 0.0000023300757253939837,
225
+ "loss": 5.546,
226
+ "step": 7500
227
+ },
228
+ {
229
+ "epoch": 2.9733881758263543,
230
+ "eval_loss": 190.01214599609375,
231
+ "eval_runtime": 571.9686,
232
+ "eval_samples_per_second": 53.134,
233
+ "eval_steps_per_second": 6.642,
234
+ "step": 7500
235
+ },
236
+ {
237
+ "epoch": 3.171614054214778,
238
+ "grad_norm": 34.228065490722656,
239
+ "learning_rate": 0.0000015628185987094685,
240
+ "loss": 5.3668,
241
+ "step": 8000
242
+ },
243
+ {
244
+ "epoch": 3.171614054214778,
245
+ "eval_loss": 194.59649658203125,
246
+ "eval_runtime": 571.773,
247
+ "eval_samples_per_second": 53.152,
248
+ "eval_steps_per_second": 6.644,
249
+ "step": 8000
250
+ },
251
+ {
252
+ "epoch": 3.3698399326032016,
253
+ "grad_norm": 30.128808975219727,
254
+ "learning_rate": 9.253760663628896e-7,
255
+ "loss": 5.338,
256
+ "step": 8500
257
+ },
258
+ {
259
+ "epoch": 3.3698399326032016,
260
+ "eval_loss": 173.66778564453125,
261
+ "eval_runtime": 569.7472,
262
+ "eval_samples_per_second": 53.341,
263
+ "eval_steps_per_second": 6.668,
264
+ "step": 8500
265
+ },
266
+ {
267
+ "epoch": 3.568065810991625,
268
+ "grad_norm": 54.32496643066406,
269
+ "learning_rate": 4.4182289604595254e-7,
270
+ "loss": 5.2961,
271
+ "step": 9000
272
+ },
273
+ {
274
+ "epoch": 3.568065810991625,
275
+ "eval_loss": 184.6847381591797,
276
+ "eval_runtime": 571.6945,
277
+ "eval_samples_per_second": 53.16,
278
+ "eval_steps_per_second": 6.645,
279
+ "step": 9000
280
+ }
281
+ ],
282
+ "logging_steps": 500,
283
+ "max_steps": 10088,
284
+ "num_input_tokens_seen": 0,
285
+ "num_train_epochs": 4,
286
+ "save_steps": 1000,
287
+ "stateful_callbacks": {
288
+ "TrainerControl": {
289
+ "args": {
290
+ "should_epoch_stop": false,
291
+ "should_evaluate": false,
292
+ "should_log": false,
293
+ "should_save": true,
294
+ "should_training_stop": false
295
+ },
296
+ "attributes": {}
297
+ }
298
+ },
299
+ "total_flos": 0,
300
+ "train_batch_size": 8,
301
+ "trial_name": null,
302
+ "trial_params": null
303
+ }