zgce commited on
Commit
f26b8c4
1 Parent(s): 28e599b

Upload 13 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "/root/autodl-tmp/Qwen2.5-14B-scipy",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 64,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "up_proj",
24
+ "o_proj",
25
+ "down_proj",
26
+ "k_proj",
27
+ "v_proj",
28
+ "gate_proj",
29
+ "q_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.9985022466300549,
3
+ "num_input_tokens_seen": 18683616,
4
+ "total_flos": 1.599323192678744e+18,
5
+ "train_loss": 2.4129163398742675,
6
+ "train_runtime": 18708.0864,
7
+ "train_samples_per_second": 0.214,
8
+ "train_steps_per_second": 0.013
9
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|im_end|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,208 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
+ "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
+ "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "model_max_length": 131072,
203
+ "pad_token": "<|endoftext|>",
204
+ "padding_side": "right",
205
+ "split_special_tokens": false,
206
+ "tokenizer_class": "Qwen2Tokenizer",
207
+ "unk_token": null
208
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 0.9985022466300549,
3
+ "num_input_tokens_seen": 18683616,
4
+ "total_flos": 1.599323192678744e+18,
5
+ "train_loss": 2.4129163398742675,
6
+ "train_runtime": 18708.0864,
7
+ "train_samples_per_second": 0.214,
8
+ "train_steps_per_second": 0.013
9
+ }
trainer_log.jsonl ADDED
@@ -0,0 +1,177 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 2, "total_steps": 250, "loss": 2.4965, "learning_rate": 0.00029995262839249497, "epoch": 0.00798801797304044, "percentage": 0.8, "elapsed_time": "0:02:45", "remaining_time": "5:41:52", "throughput": 981.74, "total_tokens": 162400}
2
+ {"current_steps": 4, "total_steps": 250, "loss": 2.4769, "learning_rate": 0.00029981054349090264, "epoch": 0.01597603594608088, "percentage": 1.6, "elapsed_time": "0:05:22", "remaining_time": "5:30:17", "throughput": 983.22, "total_tokens": 316832}
3
+ {"current_steps": 6, "total_steps": 250, "loss": 2.4562, "learning_rate": 0.00029957383503909204, "epoch": 0.023964053919121316, "percentage": 2.4, "elapsed_time": "0:07:39", "remaining_time": "5:11:29", "throughput": 993.45, "total_tokens": 456560}
4
+ {"current_steps": 8, "total_steps": 250, "loss": 2.5164, "learning_rate": 0.000299242652547195, "epoch": 0.03195207189216176, "percentage": 3.2, "elapsed_time": "0:10:15", "remaining_time": "5:10:32", "throughput": 994.0, "total_tokens": 612256}
5
+ {"current_steps": 10, "total_steps": 250, "loss": 2.4843, "learning_rate": 0.0002988172051971717, "epoch": 0.0399400898652022, "percentage": 4.0, "elapsed_time": "0:13:00", "remaining_time": "5:12:09", "throughput": 992.51, "total_tokens": 774544}
6
+ {"current_steps": 12, "total_steps": 250, "loss": 2.5155, "learning_rate": 0.0002982977617106871, "epoch": 0.04792810783824263, "percentage": 4.8, "elapsed_time": "0:15:49", "remaining_time": "5:13:56", "throughput": 989.39, "total_tokens": 939664}
7
+ {"current_steps": 14, "total_steps": 250, "loss": 2.4476, "learning_rate": 0.0002976846501793808, "epoch": 0.05591612581128307, "percentage": 5.6, "elapsed_time": "0:18:19", "remaining_time": "5:09:00", "throughput": 992.34, "total_tokens": 1091408}
8
+ {"current_steps": 16, "total_steps": 250, "loss": 2.4053, "learning_rate": 0.000296978257857637, "epoch": 0.06390414378432352, "percentage": 6.4, "elapsed_time": "0:21:01", "remaining_time": "5:07:35", "throughput": 992.14, "total_tokens": 1251968}
9
+ {"current_steps": 18, "total_steps": 250, "loss": 2.4275, "learning_rate": 0.0002961790309179866, "epoch": 0.07189216175736396, "percentage": 7.2, "elapsed_time": "0:23:45", "remaining_time": "5:06:15", "throughput": 991.55, "total_tokens": 1413632}
10
+ {"current_steps": 20, "total_steps": 250, "loss": 2.4095, "learning_rate": 0.00029528747416929463, "epoch": 0.0798801797304044, "percentage": 8.0, "elapsed_time": "0:26:12", "remaining_time": "5:01:22", "throughput": 993.12, "total_tokens": 1561616}
11
+ {"current_steps": 22, "total_steps": 250, "loss": 2.4703, "learning_rate": 0.0002943041507379128, "epoch": 0.08786819770344484, "percentage": 8.8, "elapsed_time": "0:28:31", "remaining_time": "4:55:41", "throughput": 994.26, "total_tokens": 1702048}
12
+ {"current_steps": 24, "total_steps": 250, "loss": 2.4129, "learning_rate": 0.0002932296817119964, "epoch": 0.09585621567648527, "percentage": 9.6, "elapsed_time": "0:30:55", "remaining_time": "4:51:16", "throughput": 995.28, "total_tokens": 1847168}
13
+ {"current_steps": 26, "total_steps": 250, "loss": 2.4276, "learning_rate": 0.00029206474574921165, "epoch": 0.1038442336495257, "percentage": 10.4, "elapsed_time": "0:33:33", "remaining_time": "4:49:06", "throughput": 994.66, "total_tokens": 2002624}
14
+ {"current_steps": 28, "total_steps": 250, "loss": 2.4062, "learning_rate": 0.0002908100786480811, "epoch": 0.11183225162256615, "percentage": 11.2, "elapsed_time": "0:36:18", "remaining_time": "4:47:50", "throughput": 992.99, "total_tokens": 2162992}
15
+ {"current_steps": 30, "total_steps": 250, "loss": 2.4229, "learning_rate": 0.00028946647288323766, "epoch": 0.11982026959560658, "percentage": 12.0, "elapsed_time": "0:38:38", "remaining_time": "4:43:19", "throughput": 994.15, "total_tokens": 2304544}
16
+ {"current_steps": 32, "total_steps": 250, "loss": 2.4247, "learning_rate": 0.00028803477710488055, "epoch": 0.12780828756864704, "percentage": 12.8, "elapsed_time": "0:40:56", "remaining_time": "4:38:57", "throughput": 994.88, "total_tokens": 2444336}
17
+ {"current_steps": 34, "total_steps": 250, "loss": 2.4517, "learning_rate": 0.00028651589560274934, "epoch": 0.13579630554168748, "percentage": 13.6, "elapsed_time": "0:43:33", "remaining_time": "4:36:42", "throughput": 994.44, "total_tokens": 2598848}
18
+ {"current_steps": 36, "total_steps": 250, "loss": 2.3632, "learning_rate": 0.00028491078773495564, "epoch": 0.14378432351472792, "percentage": 14.4, "elapsed_time": "0:45:59", "remaining_time": "4:33:26", "throughput": 994.78, "total_tokens": 2745520}
19
+ {"current_steps": 38, "total_steps": 250, "loss": 2.5248, "learning_rate": 0.00028322046732203165, "epoch": 0.15177234148776836, "percentage": 15.2, "elapsed_time": "0:48:50", "remaining_time": "4:32:31", "throughput": 994.03, "total_tokens": 2913392}
20
+ {"current_steps": 40, "total_steps": 250, "loss": 2.4143, "learning_rate": 0.0002814460020065795, "epoch": 0.1597603594608088, "percentage": 16.0, "elapsed_time": "0:51:10", "remaining_time": "4:28:40", "throughput": 994.48, "total_tokens": 3053600}
21
+ {"current_steps": 42, "total_steps": 250, "loss": 2.4045, "learning_rate": 0.00027958851257892526, "epoch": 0.16774837743384924, "percentage": 16.8, "elapsed_time": "0:53:33", "remaining_time": "4:25:15", "throughput": 995.5, "total_tokens": 3199248}
22
+ {"current_steps": 44, "total_steps": 250, "loss": 2.4382, "learning_rate": 0.00027764917226920377, "epoch": 0.17573639540688968, "percentage": 17.6, "elapsed_time": "0:55:35", "remaining_time": "4:20:15", "throughput": 996.71, "total_tokens": 3324464}
23
+ {"current_steps": 46, "total_steps": 250, "loss": 2.4288, "learning_rate": 0.00027562920600632127, "epoch": 0.18372441337993012, "percentage": 18.4, "elapsed_time": "0:58:20", "remaining_time": "4:18:45", "throughput": 995.47, "total_tokens": 3485088}
24
+ {"current_steps": 48, "total_steps": 250, "loss": 2.4689, "learning_rate": 0.0002735298896442641, "epoch": 0.19171243135297053, "percentage": 19.2, "elapsed_time": "1:00:49", "remaining_time": "4:16:00", "throughput": 995.96, "total_tokens": 3635152}
25
+ {"current_steps": 50, "total_steps": 250, "loss": 2.395, "learning_rate": 0.0002713525491562421, "epoch": 0.19970044932601097, "percentage": 20.0, "elapsed_time": "1:03:21", "remaining_time": "4:13:27", "throughput": 996.16, "total_tokens": 3787152}
26
+ {"current_steps": 52, "total_steps": 250, "loss": 2.3901, "learning_rate": 0.0002690985597971753, "epoch": 0.2076884672990514, "percentage": 20.8, "elapsed_time": "1:05:43", "remaining_time": "4:10:17", "throughput": 996.8, "total_tokens": 3931280}
27
+ {"current_steps": 54, "total_steps": 250, "loss": 2.3819, "learning_rate": 0.0002667693452350535, "epoch": 0.21567648527209185, "percentage": 21.6, "elapsed_time": "1:08:16", "remaining_time": "4:07:49", "throughput": 996.81, "total_tokens": 4083536}
28
+ {"current_steps": 56, "total_steps": 250, "loss": 2.444, "learning_rate": 0.0002643663766517172, "epoch": 0.2236645032451323, "percentage": 22.4, "elapsed_time": "1:10:53", "remaining_time": "4:05:36", "throughput": 996.6, "total_tokens": 4239440}
29
+ {"current_steps": 58, "total_steps": 250, "loss": 2.3407, "learning_rate": 0.00026189117181362733, "epoch": 0.23165252121817273, "percentage": 23.2, "elapsed_time": "1:12:59", "remaining_time": "4:01:36", "throughput": 997.68, "total_tokens": 4369008}
30
+ {"current_steps": 60, "total_steps": 250, "loss": 2.4137, "learning_rate": 0.0002593452941132117, "epoch": 0.23964053919121317, "percentage": 24.0, "elapsed_time": "1:15:35", "remaining_time": "3:59:23", "throughput": 997.73, "total_tokens": 4525680}
31
+ {"current_steps": 62, "total_steps": 250, "loss": 2.3727, "learning_rate": 0.0002567303515813928, "epoch": 0.2476285571642536, "percentage": 24.8, "elapsed_time": "1:18:14", "remaining_time": "3:57:15", "throughput": 997.62, "total_tokens": 4683360}
32
+ {"current_steps": 64, "total_steps": 250, "loss": 2.4705, "learning_rate": 0.0002540479958719207, "epoch": 0.2556165751372941, "percentage": 25.6, "elapsed_time": "1:20:45", "remaining_time": "3:54:43", "throughput": 997.47, "total_tokens": 4833520}
33
+ {"current_steps": 66, "total_steps": 250, "loss": 2.2845, "learning_rate": 0.00025129992121815365, "epoch": 0.2636045931103345, "percentage": 26.4, "elapsed_time": "1:23:05", "remaining_time": "3:51:37", "throughput": 997.89, "total_tokens": 4974480}
34
+ {"current_steps": 68, "total_steps": 250, "loss": 2.3124, "learning_rate": 0.00024848786336294346, "epoch": 0.27159261108337496, "percentage": 27.2, "elapsed_time": "1:25:07", "remaining_time": "3:47:49", "throughput": 998.48, "total_tokens": 5099616}
35
+ {"current_steps": 70, "total_steps": 250, "loss": 2.4011, "learning_rate": 0.0002456135984623034, "epoch": 0.2795806290564154, "percentage": 28.0, "elapsed_time": "1:27:30", "remaining_time": "3:45:00", "throughput": 998.89, "total_tokens": 5244416}
36
+ {"current_steps": 72, "total_steps": 250, "loss": 2.3147, "learning_rate": 0.00024267894196355015, "epoch": 0.28756864702945584, "percentage": 28.8, "elapsed_time": "1:29:56", "remaining_time": "3:42:20", "throughput": 999.22, "total_tokens": 5392128}
37
+ {"current_steps": 74, "total_steps": 250, "loss": 2.3653, "learning_rate": 0.00023968574745862782, "epoch": 0.2955566650024963, "percentage": 29.6, "elapsed_time": "1:32:29", "remaining_time": "3:39:57", "throughput": 999.29, "total_tokens": 5545216}
38
+ {"current_steps": 76, "total_steps": 250, "loss": 2.361, "learning_rate": 0.0002366359055133401, "epoch": 0.3035446829755367, "percentage": 30.4, "elapsed_time": "1:35:19", "remaining_time": "3:38:15", "throughput": 998.62, "total_tokens": 5712080}
39
+ {"current_steps": 78, "total_steps": 250, "loss": 2.3499, "learning_rate": 0.0002335313424732282, "epoch": 0.31153270094857716, "percentage": 31.2, "elapsed_time": "1:37:47", "remaining_time": "3:35:38", "throughput": 998.56, "total_tokens": 5858800}
40
+ {"current_steps": 80, "total_steps": 250, "loss": 2.3515, "learning_rate": 0.00023037401924684946, "epoch": 0.3195207189216176, "percentage": 32.0, "elapsed_time": "1:40:19", "remaining_time": "3:33:11", "throughput": 998.25, "total_tokens": 6008784}
41
+ {"current_steps": 82, "total_steps": 250, "loss": 2.398, "learning_rate": 0.00022716593006722594, "epoch": 0.32750873689465804, "percentage": 32.8, "elapsed_time": "1:42:45", "remaining_time": "3:30:32", "throughput": 998.34, "total_tokens": 6155536}
42
+ {"current_steps": 84, "total_steps": 250, "loss": 2.3678, "learning_rate": 0.00022390910123224373, "epoch": 0.3354967548676985, "percentage": 33.6, "elapsed_time": "1:45:17", "remaining_time": "3:28:03", "throughput": 998.33, "total_tokens": 6306576}
43
+ {"current_steps": 86, "total_steps": 250, "loss": 2.4406, "learning_rate": 0.0002206055898247999, "epoch": 0.3434847728407389, "percentage": 34.4, "elapsed_time": "1:48:04", "remaining_time": "3:26:04", "throughput": 997.91, "total_tokens": 6470448}
44
+ {"current_steps": 88, "total_steps": 250, "loss": 2.3589, "learning_rate": 0.00021725748241350486, "epoch": 0.35147279081377936, "percentage": 35.2, "elapsed_time": "1:50:29", "remaining_time": "3:23:23", "throughput": 998.17, "total_tokens": 6616864}
45
+ {"current_steps": 90, "total_steps": 250, "loss": 2.3573, "learning_rate": 0.00021386689373476087, "epoch": 0.3594608087868198, "percentage": 36.0, "elapsed_time": "1:53:25", "remaining_time": "3:21:38", "throughput": 997.21, "total_tokens": 6786448}
46
+ {"current_steps": 92, "total_steps": 250, "loss": 2.3993, "learning_rate": 0.0002104359653570494, "epoch": 0.36744882675986024, "percentage": 36.8, "elapsed_time": "1:56:18", "remaining_time": "3:19:44", "throughput": 996.64, "total_tokens": 6954720}
47
+ {"current_steps": 94, "total_steps": 250, "loss": 2.4304, "learning_rate": 0.00020696686432827017, "epoch": 0.3754368447329007, "percentage": 37.6, "elapsed_time": "1:58:36", "remaining_time": "3:16:50", "throughput": 996.85, "total_tokens": 7094208}
48
+ {"current_steps": 96, "total_steps": 250, "loss": 2.3126, "learning_rate": 0.00020346178180698758, "epoch": 0.38342486270594106, "percentage": 38.4, "elapsed_time": "2:00:44", "remaining_time": "3:13:41", "throughput": 997.64, "total_tokens": 7227440}
49
+ {"current_steps": 98, "total_steps": 250, "loss": 2.3727, "learning_rate": 0.000199922931678448, "epoch": 0.3914128806789815, "percentage": 39.2, "elapsed_time": "2:02:55", "remaining_time": "3:10:40", "throughput": 998.19, "total_tokens": 7362608}
50
+ {"current_steps": 100, "total_steps": 250, "loss": 2.3292, "learning_rate": 0.0001963525491562421, "epoch": 0.39940089865202194, "percentage": 40.0, "elapsed_time": "2:05:28", "remaining_time": "3:08:13", "throughput": 998.17, "total_tokens": 7515024}
51
+ {"current_steps": 102, "total_steps": 250, "loss": 2.3322, "learning_rate": 0.0001927528893704964, "epoch": 0.4073889166250624, "percentage": 40.8, "elapsed_time": "2:08:06", "remaining_time": "3:05:52", "throughput": 997.94, "total_tokens": 7670624}
52
+ {"current_steps": 2, "total_steps": 250, "loss": 2.5021, "learning_rate": 4.99921047320825e-05, "epoch": 0.00798801797304044, "percentage": 0.8, "elapsed_time": "0:02:44", "remaining_time": "5:40:03", "throughput": 986.97, "total_tokens": 162400}
53
+ {"current_steps": 4, "total_steps": 250, "loss": 2.5086, "learning_rate": 4.996842391515044e-05, "epoch": 0.01597603594608088, "percentage": 1.6, "elapsed_time": "0:05:21", "remaining_time": "5:29:17", "throughput": 986.19, "total_tokens": 316832}
54
+ {"current_steps": 6, "total_steps": 250, "loss": 2.4884, "learning_rate": 4.992897250651535e-05, "epoch": 0.023964053919121316, "percentage": 2.4, "elapsed_time": "0:07:38", "remaining_time": "5:10:49", "throughput": 995.56, "total_tokens": 456560}
55
+ {"current_steps": 8, "total_steps": 250, "loss": 2.5476, "learning_rate": 4.987377542453251e-05, "epoch": 0.03195207189216176, "percentage": 3.2, "elapsed_time": "0:10:15", "remaining_time": "5:10:04", "throughput": 995.49, "total_tokens": 612256}
56
+ {"current_steps": 10, "total_steps": 250, "loss": 2.5127, "learning_rate": 4.980286753286195e-05, "epoch": 0.0399400898652022, "percentage": 4.0, "elapsed_time": "0:12:59", "remaining_time": "5:11:42", "throughput": 993.92, "total_tokens": 774544}
57
+ {"current_steps": 12, "total_steps": 250, "loss": 2.5413, "learning_rate": 4.971629361844785e-05, "epoch": 0.04792810783824263, "percentage": 4.8, "elapsed_time": "0:15:48", "remaining_time": "5:13:33", "throughput": 990.58, "total_tokens": 939664}
58
+ {"current_steps": 14, "total_steps": 250, "loss": 2.4724, "learning_rate": 4.9614108363230135e-05, "epoch": 0.05591612581128307, "percentage": 5.6, "elapsed_time": "0:18:18", "remaining_time": "5:08:42", "throughput": 993.26, "total_tokens": 1091408}
59
+ {"current_steps": 16, "total_steps": 250, "loss": 2.4269, "learning_rate": 4.949637630960617e-05, "epoch": 0.06390414378432352, "percentage": 6.4, "elapsed_time": "0:21:00", "remaining_time": "5:07:17", "throughput": 993.1, "total_tokens": 1251968}
60
+ {"current_steps": 18, "total_steps": 250, "loss": 2.4497, "learning_rate": 4.9363171819664434e-05, "epoch": 0.07189216175736396, "percentage": 7.2, "elapsed_time": "0:23:44", "remaining_time": "5:05:58", "throughput": 992.44, "total_tokens": 1413632}
61
+ {"current_steps": 20, "total_steps": 250, "loss": 2.4304, "learning_rate": 4.9214579028215776e-05, "epoch": 0.0798801797304044, "percentage": 8.0, "elapsed_time": "0:26:11", "remaining_time": "5:01:08", "throughput": 993.94, "total_tokens": 1561616}
62
+ {"current_steps": 22, "total_steps": 250, "loss": 2.4962, "learning_rate": 4.905069178965215e-05, "epoch": 0.08786819770344484, "percentage": 8.8, "elapsed_time": "0:28:30", "remaining_time": "4:55:28", "throughput": 994.99, "total_tokens": 1702048}
63
+ {"current_steps": 24, "total_steps": 250, "loss": 2.4427, "learning_rate": 4.887161361866608e-05, "epoch": 0.09585621567648527, "percentage": 9.6, "elapsed_time": "0:30:54", "remaining_time": "4:51:03", "throughput": 996.03, "total_tokens": 1847168}
64
+ {"current_steps": 26, "total_steps": 250, "loss": 2.4515, "learning_rate": 4.867745762486861e-05, "epoch": 0.1038442336495257, "percentage": 10.4, "elapsed_time": "0:33:31", "remaining_time": "4:48:53", "throughput": 995.39, "total_tokens": 2002624}
65
+ {"current_steps": 28, "total_steps": 250, "loss": 2.4309, "learning_rate": 4.846834644134686e-05, "epoch": 0.11183225162256615, "percentage": 11.2, "elapsed_time": "0:36:16", "remaining_time": "4:47:38", "throughput": 993.66, "total_tokens": 2162992}
66
+ {"current_steps": 30, "total_steps": 250, "loss": 2.4502, "learning_rate": 4.8244412147206284e-05, "epoch": 0.11982026959560658, "percentage": 12.0, "elapsed_time": "0:38:36", "remaining_time": "4:43:07", "throughput": 994.83, "total_tokens": 2304544}
67
+ {"current_steps": 32, "total_steps": 250, "loss": 2.4509, "learning_rate": 4.800579618414676e-05, "epoch": 0.12780828756864704, "percentage": 12.8, "elapsed_time": "0:40:55", "remaining_time": "4:38:45", "throughput": 995.62, "total_tokens": 2444336}
68
+ {"current_steps": 34, "total_steps": 250, "loss": 2.4778, "learning_rate": 4.775264926712489e-05, "epoch": 0.13579630554168748, "percentage": 13.6, "elapsed_time": "0:43:31", "remaining_time": "4:36:32", "throughput": 995.05, "total_tokens": 2598848}
69
+ {"current_steps": 36, "total_steps": 250, "loss": 2.3898, "learning_rate": 4.7485131289159276e-05, "epoch": 0.14378432351472792, "percentage": 14.4, "elapsed_time": "0:45:58", "remaining_time": "4:33:14", "throughput": 995.46, "total_tokens": 2745520}
70
+ {"current_steps": 38, "total_steps": 250, "loss": 2.5515, "learning_rate": 4.720341122033862e-05, "epoch": 0.15177234148776836, "percentage": 15.2, "elapsed_time": "0:48:48", "remaining_time": "4:32:19", "throughput": 994.72, "total_tokens": 2913392}
71
+ {"current_steps": 40, "total_steps": 250, "loss": 2.4457, "learning_rate": 4.690766700109659e-05, "epoch": 0.1597603594608088, "percentage": 16.0, "elapsed_time": "0:51:08", "remaining_time": "4:28:28", "throughput": 995.22, "total_tokens": 3053600}
72
+ {"current_steps": 42, "total_steps": 250, "loss": 2.4336, "learning_rate": 4.659808542982088e-05, "epoch": 0.16774837743384924, "percentage": 16.8, "elapsed_time": "0:53:31", "remaining_time": "4:25:06", "throughput": 996.09, "total_tokens": 3199248}
73
+ {"current_steps": 44, "total_steps": 250, "loss": 2.4655, "learning_rate": 4.6274862044867304e-05, "epoch": 0.17573639540688968, "percentage": 17.6, "elapsed_time": "0:55:32", "remaining_time": "4:20:03", "throughput": 997.5, "total_tokens": 3324464}
74
+ {"current_steps": 46, "total_steps": 250, "loss": 2.4566, "learning_rate": 4.593820100105355e-05, "epoch": 0.18372441337993012, "percentage": 18.4, "elapsed_time": "0:58:18", "remaining_time": "4:18:34", "throughput": 996.21, "total_tokens": 3485088}
75
+ {"current_steps": 48, "total_steps": 250, "loss": 2.4912, "learning_rate": 4.558831494071069e-05, "epoch": 0.19171243135297053, "percentage": 19.2, "elapsed_time": "1:00:47", "remaining_time": "4:15:48", "throughput": 996.69, "total_tokens": 3635152}
76
+ {"current_steps": 50, "total_steps": 250, "loss": 2.4185, "learning_rate": 4.522542485937369e-05, "epoch": 0.19970044932601097, "percentage": 20.0, "elapsed_time": "1:03:19", "remaining_time": "4:13:16", "throughput": 996.88, "total_tokens": 3787152}
77
+ {"current_steps": 52, "total_steps": 250, "loss": 2.4221, "learning_rate": 4.484975996619589e-05, "epoch": 0.2076884672990514, "percentage": 20.8, "elapsed_time": "1:05:41", "remaining_time": "4:10:06", "throughput": 997.5, "total_tokens": 3931280}
78
+ {"current_steps": 54, "total_steps": 250, "loss": 2.4091, "learning_rate": 4.4461557539175594e-05, "epoch": 0.21567648527209185, "percentage": 21.6, "elapsed_time": "1:08:13", "remaining_time": "4:07:38", "throughput": 997.5, "total_tokens": 4083536}
79
+ {"current_steps": 56, "total_steps": 250, "loss": 2.4697, "learning_rate": 4.40610627752862e-05, "epoch": 0.2236645032451323, "percentage": 22.4, "elapsed_time": "1:10:50", "remaining_time": "4:05:26", "throughput": 997.29, "total_tokens": 4239440}
80
+ {"current_steps": 58, "total_steps": 250, "loss": 2.3675, "learning_rate": 4.3648528635604556e-05, "epoch": 0.23165252121817273, "percentage": 23.2, "elapsed_time": "1:12:56", "remaining_time": "4:01:26", "throughput": 998.36, "total_tokens": 4369008}
81
+ {"current_steps": 60, "total_steps": 250, "loss": 2.4421, "learning_rate": 4.3224215685535294e-05, "epoch": 0.23964053919121317, "percentage": 24.0, "elapsed_time": "1:15:33", "remaining_time": "3:59:14", "throughput": 998.36, "total_tokens": 4525680}
82
+ {"current_steps": 62, "total_steps": 250, "loss": 2.3973, "learning_rate": 4.278839193023214e-05, "epoch": 0.2476285571642536, "percentage": 24.8, "elapsed_time": "1:18:11", "remaining_time": "3:57:06", "throughput": 998.23, "total_tokens": 4683360}
83
+ {"current_steps": 64, "total_steps": 250, "loss": 2.5022, "learning_rate": 4.234133264532012e-05, "epoch": 0.2556165751372941, "percentage": 25.6, "elapsed_time": "1:20:42", "remaining_time": "3:54:34", "throughput": 998.06, "total_tokens": 4833520}
84
+ {"current_steps": 66, "total_steps": 250, "loss": 2.3089, "learning_rate": 4.188332020302561e-05, "epoch": 0.2636045931103345, "percentage": 26.4, "elapsed_time": "1:23:02", "remaining_time": "3:51:29", "throughput": 998.46, "total_tokens": 4974480}
85
+ {"current_steps": 68, "total_steps": 250, "loss": 2.3378, "learning_rate": 4.1414643893823914e-05, "epoch": 0.27159261108337496, "percentage": 27.2, "elapsed_time": "1:25:04", "remaining_time": "3:47:42", "throughput": 999.02, "total_tokens": 5099616}
86
+ {"current_steps": 70, "total_steps": 250, "loss": 2.4277, "learning_rate": 4.093559974371725e-05, "epoch": 0.2795806290564154, "percentage": 28.0, "elapsed_time": "1:27:27", "remaining_time": "3:44:53", "throughput": 999.43, "total_tokens": 5244416}
87
+ {"current_steps": 72, "total_steps": 250, "loss": 2.3468, "learning_rate": 4.044649032725836e-05, "epoch": 0.28756864702945584, "percentage": 28.8, "elapsed_time": "1:29:53", "remaining_time": "3:42:13", "throughput": 999.76, "total_tokens": 5392128}
88
+ {"current_steps": 74, "total_steps": 250, "loss": 2.3907, "learning_rate": 3.9947624576437975e-05, "epoch": 0.2955566650024963, "percentage": 29.6, "elapsed_time": "1:32:26", "remaining_time": "3:39:51", "throughput": 999.79, "total_tokens": 5545216}
89
+ {"current_steps": 76, "total_steps": 250, "loss": 2.3878, "learning_rate": 3.943931758555669e-05, "epoch": 0.3035446829755367, "percentage": 30.4, "elapsed_time": "1:35:17", "remaining_time": "3:38:09", "throughput": 999.1, "total_tokens": 5712080}
90
+ {"current_steps": 78, "total_steps": 250, "loss": 2.3853, "learning_rate": 3.8921890412204705e-05, "epoch": 0.31153270094857716, "percentage": 31.2, "elapsed_time": "1:37:44", "remaining_time": "3:35:31", "throughput": 999.04, "total_tokens": 5858800}
91
+ {"current_steps": 80, "total_steps": 250, "loss": 2.3765, "learning_rate": 3.8395669874474915e-05, "epoch": 0.3195207189216176, "percentage": 32.0, "elapsed_time": "1:40:15", "remaining_time": "3:33:03", "throughput": 998.81, "total_tokens": 6008784}
92
+ {"current_steps": 82, "total_steps": 250, "loss": 2.4277, "learning_rate": 3.786098834453766e-05, "epoch": 0.32750873689465804, "percentage": 32.8, "elapsed_time": "1:42:43", "remaining_time": "3:30:26", "throughput": 998.78, "total_tokens": 6155536}
93
+ {"current_steps": 84, "total_steps": 250, "loss": 2.3935, "learning_rate": 3.731818353870729e-05, "epoch": 0.3354967548676985, "percentage": 33.6, "elapsed_time": "1:45:14", "remaining_time": "3:27:58", "throughput": 998.76, "total_tokens": 6306576}
94
+ {"current_steps": 86, "total_steps": 250, "loss": 2.4679, "learning_rate": 3.6767598304133324e-05, "epoch": 0.3434847728407389, "percentage": 34.4, "elapsed_time": "1:48:00", "remaining_time": "3:25:58", "throughput": 998.38, "total_tokens": 6470448}
95
+ {"current_steps": 88, "total_steps": 250, "loss": 2.3868, "learning_rate": 3.6209580402250815e-05, "epoch": 0.35147279081377936, "percentage": 35.2, "elapsed_time": "1:50:25", "remaining_time": "3:23:17", "throughput": 998.63, "total_tokens": 6616864}
96
+ {"current_steps": 90, "total_steps": 250, "loss": 2.3859, "learning_rate": 3.564448228912682e-05, "epoch": 0.3594608087868198, "percentage": 36.0, "elapsed_time": "1:53:22", "remaining_time": "3:21:33", "throughput": 997.62, "total_tokens": 6786448}
97
+ {"current_steps": 92, "total_steps": 250, "loss": 2.4216, "learning_rate": 3.507266089284157e-05, "epoch": 0.36744882675986024, "percentage": 36.8, "elapsed_time": "1:56:14", "remaining_time": "3:19:38", "throughput": 997.13, "total_tokens": 6954720}
98
+ {"current_steps": 94, "total_steps": 250, "loss": 2.4611, "learning_rate": 3.4494477388045035e-05, "epoch": 0.3754368447329007, "percentage": 37.6, "elapsed_time": "1:58:33", "remaining_time": "3:16:44", "throughput": 997.32, "total_tokens": 7094208}
99
+ {"current_steps": 96, "total_steps": 250, "loss": 2.3384, "learning_rate": 3.3910296967831266e-05, "epoch": 0.38342486270594106, "percentage": 38.4, "elapsed_time": "2:00:41", "remaining_time": "3:13:36", "throughput": 998.11, "total_tokens": 7227440}
100
+ {"current_steps": 98, "total_steps": 250, "loss": 2.4074, "learning_rate": 3.332048861307467e-05, "epoch": 0.3914128806789815, "percentage": 39.2, "elapsed_time": "2:02:52", "remaining_time": "3:10:34", "throughput": 998.66, "total_tokens": 7362608}
101
+ {"current_steps": 100, "total_steps": 250, "loss": 2.3572, "learning_rate": 3.272542485937369e-05, "epoch": 0.39940089865202194, "percentage": 40.0, "elapsed_time": "2:05:25", "remaining_time": "3:08:07", "throughput": 998.63, "total_tokens": 7515024}
102
+ {"current_steps": 102, "total_steps": 250, "loss": 2.3579, "learning_rate": 3.21254815617494e-05, "epoch": 0.4073889166250624, "percentage": 40.8, "elapsed_time": "2:08:02", "remaining_time": "3:05:47", "throughput": 998.41, "total_tokens": 7670624}
103
+ {"current_steps": 104, "total_steps": 250, "loss": 2.3728, "learning_rate": 3.152103765724743e-05, "epoch": 0.4153769345981028, "percentage": 41.6, "elapsed_time": "2:10:28", "remaining_time": "3:03:09", "throughput": 998.66, "total_tokens": 7817648}
104
+ {"current_steps": 106, "total_steps": 250, "loss": 2.2712, "learning_rate": 3.091247492559312e-05, "epoch": 0.42336495257114326, "percentage": 42.4, "elapsed_time": "2:12:58", "remaining_time": "3:00:38", "throughput": 998.75, "total_tokens": 7968432}
105
+ {"current_steps": 108, "total_steps": 250, "loss": 2.4054, "learning_rate": 3.0300177748051373e-05, "epoch": 0.4313529705441837, "percentage": 43.2, "elapsed_time": "2:15:40", "remaining_time": "2:58:23", "throughput": 998.38, "total_tokens": 8127312}
106
+ {"current_steps": 110, "total_steps": 250, "loss": 2.3966, "learning_rate": 2.9684532864643122e-05, "epoch": 0.43934098851722414, "percentage": 44.0, "elapsed_time": "2:18:08", "remaining_time": "2:55:49", "throughput": 998.32, "total_tokens": 8275008}
107
+ {"current_steps": 112, "total_steps": 250, "loss": 2.4478, "learning_rate": 2.9065929129872094e-05, "epoch": 0.4473290064902646, "percentage": 44.8, "elapsed_time": "2:20:40", "remaining_time": "2:53:19", "throughput": 998.23, "total_tokens": 8425536}
108
+ {"current_steps": 114, "total_steps": 250, "loss": 2.4386, "learning_rate": 2.844475726711595e-05, "epoch": 0.455317024463305, "percentage": 45.6, "elapsed_time": "2:23:09", "remaining_time": "2:50:46", "throughput": 998.21, "total_tokens": 8573776}
109
+ {"current_steps": 116, "total_steps": 250, "loss": 2.4176, "learning_rate": 2.782140962183704e-05, "epoch": 0.46330504243634546, "percentage": 46.4, "elapsed_time": "2:25:44", "remaining_time": "2:48:21", "throughput": 998.29, "total_tokens": 8729200}
110
+ {"current_steps": 118, "total_steps": 250, "loss": 2.3505, "learning_rate": 2.7196279913768584e-05, "epoch": 0.4712930604093859, "percentage": 47.2, "elapsed_time": "2:28:21", "remaining_time": "2:45:57", "throughput": 998.13, "total_tokens": 8884720}
111
+ {"current_steps": 120, "total_steps": 250, "loss": 2.3867, "learning_rate": 2.656976298823284e-05, "epoch": 0.47928107838242634, "percentage": 48.0, "elapsed_time": "2:30:54", "remaining_time": "2:43:28", "throughput": 997.88, "total_tokens": 9035088}
112
+ {"current_steps": 122, "total_steps": 250, "loss": 2.4321, "learning_rate": 2.594225456674837e-05, "epoch": 0.4872690963554668, "percentage": 48.8, "elapsed_time": "2:33:28", "remaining_time": "2:41:00", "throughput": 997.87, "total_tokens": 9188528}
113
+ {"current_steps": 124, "total_steps": 250, "loss": 2.4392, "learning_rate": 2.531415099708382e-05, "epoch": 0.4952571143285072, "percentage": 49.6, "elapsed_time": "2:36:15", "remaining_time": "2:38:47", "throughput": 997.54, "total_tokens": 9352752}
114
+ {"current_steps": 126, "total_steps": 250, "loss": 2.4005, "learning_rate": 2.4685849002916183e-05, "epoch": 0.5032451323015477, "percentage": 50.4, "elapsed_time": "2:38:40", "remaining_time": "2:36:09", "throughput": 997.62, "total_tokens": 9497840}
115
+ {"current_steps": 128, "total_steps": 250, "loss": 2.309, "learning_rate": 2.4057745433251635e-05, "epoch": 0.5112331502745882, "percentage": 51.2, "elapsed_time": "2:41:16", "remaining_time": "2:33:43", "throughput": 997.73, "total_tokens": 9654816}
116
+ {"current_steps": 130, "total_steps": 250, "loss": 2.522, "learning_rate": 2.3430237011767167e-05, "epoch": 0.5192211682476285, "percentage": 52.0, "elapsed_time": "2:43:59", "remaining_time": "2:31:22", "throughput": 997.5, "total_tokens": 9814944}
117
+ {"current_steps": 132, "total_steps": 250, "loss": 2.4151, "learning_rate": 2.280372008623142e-05, "epoch": 0.527209186220669, "percentage": 52.8, "elapsed_time": "2:46:14", "remaining_time": "2:28:36", "throughput": 997.77, "total_tokens": 9951952}
118
+ {"current_steps": 134, "total_steps": 250, "loss": 2.3723, "learning_rate": 2.217859037816296e-05, "epoch": 0.5351972041937094, "percentage": 53.6, "elapsed_time": "2:48:58", "remaining_time": "2:26:16", "throughput": 997.45, "total_tokens": 10112272}
119
+ {"current_steps": 136, "total_steps": 250, "loss": 2.416, "learning_rate": 2.155524273288405e-05, "epoch": 0.5431852221667499, "percentage": 54.4, "elapsed_time": "2:51:51", "remaining_time": "2:24:03", "throughput": 997.06, "total_tokens": 10281344}
120
+ {"current_steps": 138, "total_steps": 250, "loss": 2.5335, "learning_rate": 2.0934070870127912e-05, "epoch": 0.5511732401397903, "percentage": 55.2, "elapsed_time": "2:53:55", "remaining_time": "2:21:09", "throughput": 997.52, "total_tokens": 10409584}
121
+ {"current_steps": 140, "total_steps": 250, "loss": 2.4946, "learning_rate": 2.031546713535688e-05, "epoch": 0.5591612581128308, "percentage": 56.0, "elapsed_time": "2:56:43", "remaining_time": "2:18:51", "throughput": 997.17, "total_tokens": 10573840}
122
+ {"current_steps": 142, "total_steps": 250, "loss": 2.3169, "learning_rate": 1.969982225194864e-05, "epoch": 0.5671492760858712, "percentage": 56.8, "elapsed_time": "2:59:20", "remaining_time": "2:16:23", "throughput": 997.15, "total_tokens": 10729760}
123
+ {"current_steps": 144, "total_steps": 250, "loss": 2.3797, "learning_rate": 1.908752507440689e-05, "epoch": 0.5751372940589117, "percentage": 57.6, "elapsed_time": "3:01:52", "remaining_time": "2:13:52", "throughput": 997.18, "total_tokens": 10881664}
124
+ {"current_steps": 146, "total_steps": 250, "loss": 2.4498, "learning_rate": 1.8478962342752583e-05, "epoch": 0.5831253120319521, "percentage": 58.4, "elapsed_time": "3:04:25", "remaining_time": "2:11:22", "throughput": 997.2, "total_tokens": 11034352}
125
+ {"current_steps": 148, "total_steps": 250, "loss": 2.3818, "learning_rate": 1.7874518438250597e-05, "epoch": 0.5911133300049926, "percentage": 59.2, "elapsed_time": "3:06:45", "remaining_time": "2:08:42", "throughput": 997.38, "total_tokens": 11175952}
126
+ {"current_steps": 150, "total_steps": 250, "loss": 2.3818, "learning_rate": 1.7274575140626318e-05, "epoch": 0.5991013479780329, "percentage": 60.0, "elapsed_time": "3:09:00", "remaining_time": "2:06:00", "throughput": 997.69, "total_tokens": 11314112}
127
+ {"current_steps": 152, "total_steps": 250, "loss": 2.4073, "learning_rate": 1.6679511386925337e-05, "epoch": 0.6070893659510734, "percentage": 60.8, "elapsed_time": "3:11:51", "remaining_time": "2:03:42", "throughput": 997.38, "total_tokens": 11481792}
128
+ {"current_steps": 154, "total_steps": 250, "loss": 2.3467, "learning_rate": 1.6089703032168733e-05, "epoch": 0.6150773839241138, "percentage": 61.6, "elapsed_time": "3:14:44", "remaining_time": "2:01:23", "throughput": 996.97, "total_tokens": 11649040}
129
+ {"current_steps": 156, "total_steps": 250, "loss": 2.4047, "learning_rate": 1.5505522611954975e-05, "epoch": 0.6230654018971543, "percentage": 62.4, "elapsed_time": "3:17:07", "remaining_time": "1:58:46", "throughput": 996.99, "total_tokens": 11791488}
130
+ {"current_steps": 158, "total_steps": 250, "loss": 2.2998, "learning_rate": 1.4927339107158437e-05, "epoch": 0.6310534198701947, "percentage": 63.2, "elapsed_time": "3:19:38", "remaining_time": "1:56:14", "throughput": 996.81, "total_tokens": 11940144}
131
+ {"current_steps": 160, "total_steps": 250, "loss": 2.383, "learning_rate": 1.4355517710873184e-05, "epoch": 0.6390414378432352, "percentage": 64.0, "elapsed_time": "3:22:16", "remaining_time": "1:53:46", "throughput": 996.79, "total_tokens": 12097648}
132
+ {"current_steps": 162, "total_steps": 250, "loss": 2.3525, "learning_rate": 1.3790419597749199e-05, "epoch": 0.6470294558162756, "percentage": 64.8, "elapsed_time": "3:24:53", "remaining_time": "1:51:17", "throughput": 996.74, "total_tokens": 12253152}
133
+ {"current_steps": 164, "total_steps": 250, "loss": 2.4133, "learning_rate": 1.3232401695866687e-05, "epoch": 0.6550174737893161, "percentage": 65.6, "elapsed_time": "3:27:33", "remaining_time": "1:48:50", "throughput": 996.51, "total_tokens": 12410080}
134
+ {"current_steps": 166, "total_steps": 250, "loss": 2.3677, "learning_rate": 1.2681816461292715e-05, "epoch": 0.6630054917623565, "percentage": 66.4, "elapsed_time": "3:29:44", "remaining_time": "1:46:08", "throughput": 996.82, "total_tokens": 12544848}
135
+ {"current_steps": 168, "total_steps": 250, "loss": 2.4215, "learning_rate": 1.2139011655462337e-05, "epoch": 0.670993509735397, "percentage": 67.2, "elapsed_time": "3:32:22", "remaining_time": "1:43:39", "throughput": 996.77, "total_tokens": 12701808}
136
+ {"current_steps": 170, "total_steps": 250, "loss": 2.3162, "learning_rate": 1.1604330125525079e-05, "epoch": 0.6789815277084373, "percentage": 68.0, "elapsed_time": "3:34:48", "remaining_time": "1:41:05", "throughput": 996.68, "total_tokens": 12845840}
137
+ {"current_steps": 172, "total_steps": 250, "loss": 2.4129, "learning_rate": 1.107810958779531e-05, "epoch": 0.6869695456814778, "percentage": 68.8, "elapsed_time": "3:37:09", "remaining_time": "1:38:28", "throughput": 996.83, "total_tokens": 12988144}
138
+ {"current_steps": 174, "total_steps": 250, "loss": 2.4046, "learning_rate": 1.0560682414443315e-05, "epoch": 0.6949575636545182, "percentage": 69.6, "elapsed_time": "3:39:29", "remaining_time": "1:35:51", "throughput": 997.02, "total_tokens": 13129808}
139
+ {"current_steps": 176, "total_steps": 250, "loss": 2.3596, "learning_rate": 1.0052375423562038e-05, "epoch": 0.7029455816275587, "percentage": 70.4, "elapsed_time": "3:41:47", "remaining_time": "1:33:15", "throughput": 997.1, "total_tokens": 13269264}
140
+ {"current_steps": 178, "total_steps": 250, "loss": 2.3047, "learning_rate": 9.553509672741645e-06, "epoch": 0.7109335996005991, "percentage": 71.2, "elapsed_time": "3:44:16", "remaining_time": "1:30:43", "throughput": 997.07, "total_tokens": 13417184}
141
+ {"current_steps": 180, "total_steps": 250, "loss": 2.4177, "learning_rate": 9.064400256282757e-06, "epoch": 0.7189216175736396, "percentage": 72.0, "elapsed_time": "3:46:41", "remaining_time": "1:28:09", "throughput": 997.03, "total_tokens": 13561008}
142
+ {"current_steps": 182, "total_steps": 250, "loss": 2.4451, "learning_rate": 8.585356106176094e-06, "epoch": 0.72690963554668, "percentage": 72.8, "elapsed_time": "3:49:05", "remaining_time": "1:25:35", "throughput": 997.19, "total_tokens": 13706976}
143
+ {"current_steps": 184, "total_steps": 250, "loss": 2.3997, "learning_rate": 8.116679796974388e-06, "epoch": 0.7348976535197205, "percentage": 73.6, "elapsed_time": "3:51:09", "remaining_time": "1:22:55", "throughput": 997.55, "total_tokens": 13835936}
144
+ {"current_steps": 186, "total_steps": 250, "loss": 2.412, "learning_rate": 7.65866735467988e-06, "epoch": 0.7428856714927609, "percentage": 74.4, "elapsed_time": "3:53:11", "remaining_time": "1:20:14", "throughput": 997.91, "total_tokens": 13962144}
145
+ {"current_steps": 188, "total_steps": 250, "loss": 2.4054, "learning_rate": 7.211608069767867e-06, "epoch": 0.7508736894658014, "percentage": 75.2, "elapsed_time": "3:56:02", "remaining_time": "1:17:50", "throughput": 997.61, "total_tokens": 14128800}
146
+ {"current_steps": 190, "total_steps": 250, "loss": 2.4417, "learning_rate": 6.775784314464717e-06, "epoch": 0.7588617074388417, "percentage": 76.0, "elapsed_time": "3:58:32", "remaining_time": "1:15:19", "throughput": 997.62, "total_tokens": 14278288}
147
+ {"current_steps": 192, "total_steps": 250, "loss": 2.3358, "learning_rate": 6.3514713643954475e-06, "epoch": 0.7668497254118821, "percentage": 76.8, "elapsed_time": "4:01:04", "remaining_time": "1:12:49", "throughput": 997.64, "total_tokens": 14429936}
148
+ {"current_steps": 194, "total_steps": 250, "loss": 2.5089, "learning_rate": 5.9389372247138e-06, "epoch": 0.7748377433849226, "percentage": 77.6, "elapsed_time": "4:03:18", "remaining_time": "1:10:14", "throughput": 997.85, "total_tokens": 14567552}
149
+ {"current_steps": 196, "total_steps": 250, "loss": 2.4246, "learning_rate": 5.538442460824417e-06, "epoch": 0.782825761357963, "percentage": 78.4, "elapsed_time": "4:05:33", "remaining_time": "1:07:39", "throughput": 998.01, "total_tokens": 14704528}
150
+ {"current_steps": 198, "total_steps": 250, "loss": 2.369, "learning_rate": 5.150240033804116e-06, "epoch": 0.7908137793310035, "percentage": 79.2, "elapsed_time": "4:07:52", "remaining_time": "1:05:05", "throughput": 998.23, "total_tokens": 14846192}
151
+ {"current_steps": 200, "total_steps": 250, "loss": 2.4204, "learning_rate": 4.7745751406263165e-06, "epoch": 0.7988017973040439, "percentage": 80.0, "elapsed_time": "4:10:19", "remaining_time": "1:02:34", "throughput": 998.28, "total_tokens": 14993936}
152
+ {"current_steps": 202, "total_steps": 250, "loss": 2.3141, "learning_rate": 4.411685059289314e-06, "epoch": 0.8067898152770844, "percentage": 80.8, "elapsed_time": "4:12:39", "remaining_time": "1:00:02", "throughput": 998.43, "total_tokens": 15136240}
153
+ {"current_steps": 204, "total_steps": 250, "loss": 2.3841, "learning_rate": 4.061798998946459e-06, "epoch": 0.8147778332501248, "percentage": 81.6, "elapsed_time": "4:14:57", "remaining_time": "0:57:29", "throughput": 998.54, "total_tokens": 15274672}
154
+ {"current_steps": 206, "total_steps": 250, "loss": 2.3653, "learning_rate": 3.725137955132707e-06, "epoch": 0.8227658512231653, "percentage": 82.4, "elapsed_time": "4:17:18", "remaining_time": "0:54:57", "throughput": 998.67, "total_tokens": 15417920}
155
+ {"current_steps": 208, "total_steps": 250, "loss": 2.4382, "learning_rate": 3.4019145701791184e-06, "epoch": 0.8307538691962056, "percentage": 83.2, "elapsed_time": "4:19:37", "remaining_time": "0:52:25", "throughput": 998.73, "total_tokens": 15557920}
156
+ {"current_steps": 210, "total_steps": 250, "loss": 2.4455, "learning_rate": 3.092332998903416e-06, "epoch": 0.8387418871692461, "percentage": 84.0, "elapsed_time": "4:21:42", "remaining_time": "0:49:50", "throughput": 999.09, "total_tokens": 15687968}
157
+ {"current_steps": 212, "total_steps": 250, "loss": 2.4322, "learning_rate": 2.7965887796613884e-06, "epoch": 0.8467299051422865, "percentage": 84.8, "elapsed_time": "4:24:17", "remaining_time": "0:47:22", "throughput": 999.1, "total_tokens": 15843072}
158
+ {"current_steps": 214, "total_steps": 250, "loss": 2.4433, "learning_rate": 2.514868710840723e-06, "epoch": 0.854717923115327, "percentage": 85.6, "elapsed_time": "4:26:59", "remaining_time": "0:44:54", "throughput": 998.92, "total_tokens": 16002112}
159
+ {"current_steps": 216, "total_steps": 250, "loss": 2.4399, "learning_rate": 2.2473507328751086e-06, "epoch": 0.8627059410883674, "percentage": 86.4, "elapsed_time": "4:29:08", "remaining_time": "0:42:21", "throughput": 999.17, "total_tokens": 16135328}
160
+ {"current_steps": 218, "total_steps": 250, "loss": 2.378, "learning_rate": 1.9942038158532407e-06, "epoch": 0.8706939590614079, "percentage": 87.2, "elapsed_time": "4:31:49", "remaining_time": "0:39:54", "throughput": 999.01, "total_tokens": 16293008}
161
+ {"current_steps": 220, "total_steps": 250, "loss": 2.3874, "learning_rate": 1.7555878527937164e-06, "epoch": 0.8786819770344483, "percentage": 88.0, "elapsed_time": "4:34:16", "remaining_time": "0:37:24", "throughput": 999.09, "total_tokens": 16441024}
162
+ {"current_steps": 222, "total_steps": 250, "loss": 2.4028, "learning_rate": 1.5316535586531483e-06, "epoch": 0.8866699950074888, "percentage": 88.8, "elapsed_time": "4:36:40", "remaining_time": "0:34:53", "throughput": 999.21, "total_tokens": 16587616}
163
+ {"current_steps": 224, "total_steps": 250, "loss": 2.3372, "learning_rate": 1.3225423751313942e-06, "epoch": 0.8946580129805292, "percentage": 89.6, "elapsed_time": "4:39:02", "remaining_time": "0:32:23", "throughput": 999.27, "total_tokens": 16730800}
164
+ {"current_steps": 226, "total_steps": 250, "loss": 2.4288, "learning_rate": 1.1283863813339263e-06, "epoch": 0.9026460309535697, "percentage": 90.4, "elapsed_time": "4:41:37", "remaining_time": "0:29:54", "throughput": 999.31, "total_tokens": 16886144}
165
+ {"current_steps": 228, "total_steps": 250, "loss": 2.3483, "learning_rate": 9.493082103478517e-07, "epoch": 0.91063404892661, "percentage": 91.2, "elapsed_time": "4:44:13", "remaining_time": "0:27:25", "throughput": 999.16, "total_tokens": 17038816}
166
+ {"current_steps": 230, "total_steps": 250, "loss": 2.4648, "learning_rate": 7.854209717842231e-07, "epoch": 0.9186220668996505, "percentage": 92.0, "elapsed_time": "4:46:23", "remaining_time": "0:24:54", "throughput": 999.34, "total_tokens": 17171792}
167
+ {"current_steps": 232, "total_steps": 250, "loss": 2.4327, "learning_rate": 6.368281803355691e-07, "epoch": 0.9266100848726909, "percentage": 92.8, "elapsed_time": "4:48:56", "remaining_time": "0:22:25", "throughput": 999.34, "total_tokens": 17324752}
168
+ {"current_steps": 234, "total_steps": 250, "loss": 2.3037, "learning_rate": 5.036236903938285e-07, "epoch": 0.9345981028457314, "percentage": 93.6, "elapsed_time": "4:51:39", "remaining_time": "0:19:56", "throughput": 999.16, "total_tokens": 17485168}
169
+ {"current_steps": 236, "total_steps": 250, "loss": 2.4094, "learning_rate": 3.8589163676986674e-07, "epoch": 0.9425861208187718, "percentage": 94.4, "elapsed_time": "4:54:08", "remaining_time": "0:17:26", "throughput": 999.11, "total_tokens": 17632832}
170
+ {"current_steps": 238, "total_steps": 250, "loss": 2.3348, "learning_rate": 2.8370638155215123e-07, "epoch": 0.9505741387918123, "percentage": 95.2, "elapsed_time": "4:56:39", "remaining_time": "0:14:57", "throughput": 999.03, "total_tokens": 17781744}
171
+ {"current_steps": 240, "total_steps": 250, "loss": 2.3979, "learning_rate": 1.9713246713805588e-07, "epoch": 0.9585621567648527, "percentage": 96.0, "elapsed_time": "4:59:44", "remaining_time": "0:12:29", "throughput": 998.58, "total_tokens": 17959392}
172
+ {"current_steps": 242, "total_steps": 250, "loss": 2.346, "learning_rate": 1.2622457546749567e-07, "epoch": 0.9665501747378932, "percentage": 96.8, "elapsed_time": "5:02:29", "remaining_time": "0:09:59", "throughput": 998.47, "total_tokens": 18121744}
173
+ {"current_steps": 244, "total_steps": 250, "loss": 2.4486, "learning_rate": 7.102749348465165e-08, "epoch": 0.9745381927109336, "percentage": 97.6, "elapsed_time": "5:05:03", "remaining_time": "0:07:30", "throughput": 998.39, "total_tokens": 18273856}
174
+ {"current_steps": 246, "total_steps": 250, "loss": 2.3959, "learning_rate": 3.157608484956332e-08, "epoch": 0.982526210683974, "percentage": 98.4, "elapsed_time": "5:07:14", "remaining_time": "0:04:59", "throughput": 998.56, "total_tokens": 18407456}
175
+ {"current_steps": 248, "total_steps": 250, "loss": 2.4287, "learning_rate": 7.895267917501504e-09, "epoch": 0.9905142286570144, "percentage": 99.2, "elapsed_time": "5:09:43", "remaining_time": "0:02:29", "throughput": 998.52, "total_tokens": 18555776}
176
+ {"current_steps": 250, "total_steps": 250, "loss": 2.4911, "learning_rate": 0.0, "epoch": 0.9985022466300549, "percentage": 100.0, "elapsed_time": "5:11:44", "remaining_time": "0:00:00", "throughput": 998.87, "total_tokens": 18683616}
177
+ {"current_steps": 250, "total_steps": 250, "epoch": 0.9985022466300549, "percentage": 100.0, "elapsed_time": "5:11:48", "remaining_time": "0:00:00", "throughput": 998.69, "total_tokens": 18683616}
trainer_state.json ADDED
@@ -0,0 +1,1043 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 0.9985022466300549,
5
+ "eval_steps": 500,
6
+ "global_step": 250,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.00798801797304044,
13
+ "grad_norm": 0.02341790311038494,
14
+ "learning_rate": 4.99921047320825e-05,
15
+ "loss": 2.5021,
16
+ "num_input_tokens_seen": 162400,
17
+ "step": 2
18
+ },
19
+ {
20
+ "epoch": 0.01597603594608088,
21
+ "grad_norm": 0.030116591602563858,
22
+ "learning_rate": 4.996842391515044e-05,
23
+ "loss": 2.5086,
24
+ "num_input_tokens_seen": 316832,
25
+ "step": 4
26
+ },
27
+ {
28
+ "epoch": 0.023964053919121316,
29
+ "grad_norm": 0.03529619425535202,
30
+ "learning_rate": 4.992897250651535e-05,
31
+ "loss": 2.4884,
32
+ "num_input_tokens_seen": 456560,
33
+ "step": 6
34
+ },
35
+ {
36
+ "epoch": 0.03195207189216176,
37
+ "grad_norm": 0.028305748477578163,
38
+ "learning_rate": 4.987377542453251e-05,
39
+ "loss": 2.5476,
40
+ "num_input_tokens_seen": 612256,
41
+ "step": 8
42
+ },
43
+ {
44
+ "epoch": 0.0399400898652022,
45
+ "grad_norm": 0.02712586335837841,
46
+ "learning_rate": 4.980286753286195e-05,
47
+ "loss": 2.5127,
48
+ "num_input_tokens_seen": 774544,
49
+ "step": 10
50
+ },
51
+ {
52
+ "epoch": 0.04792810783824263,
53
+ "grad_norm": 0.029004713520407677,
54
+ "learning_rate": 4.971629361844785e-05,
55
+ "loss": 2.5413,
56
+ "num_input_tokens_seen": 939664,
57
+ "step": 12
58
+ },
59
+ {
60
+ "epoch": 0.05591612581128307,
61
+ "grad_norm": 0.02819664776325226,
62
+ "learning_rate": 4.9614108363230135e-05,
63
+ "loss": 2.4724,
64
+ "num_input_tokens_seen": 1091408,
65
+ "step": 14
66
+ },
67
+ {
68
+ "epoch": 0.06390414378432352,
69
+ "grad_norm": 0.025089334696531296,
70
+ "learning_rate": 4.949637630960617e-05,
71
+ "loss": 2.4269,
72
+ "num_input_tokens_seen": 1251968,
73
+ "step": 16
74
+ },
75
+ {
76
+ "epoch": 0.07189216175736396,
77
+ "grad_norm": 0.027914991602301598,
78
+ "learning_rate": 4.9363171819664434e-05,
79
+ "loss": 2.4497,
80
+ "num_input_tokens_seen": 1413632,
81
+ "step": 18
82
+ },
83
+ {
84
+ "epoch": 0.0798801797304044,
85
+ "grad_norm": 0.029010389000177383,
86
+ "learning_rate": 4.9214579028215776e-05,
87
+ "loss": 2.4304,
88
+ "num_input_tokens_seen": 1561616,
89
+ "step": 20
90
+ },
91
+ {
92
+ "epoch": 0.08786819770344484,
93
+ "grad_norm": 0.032729391008615494,
94
+ "learning_rate": 4.905069178965215e-05,
95
+ "loss": 2.4962,
96
+ "num_input_tokens_seen": 1702048,
97
+ "step": 22
98
+ },
99
+ {
100
+ "epoch": 0.09585621567648527,
101
+ "grad_norm": 0.034885670989751816,
102
+ "learning_rate": 4.887161361866608e-05,
103
+ "loss": 2.4427,
104
+ "num_input_tokens_seen": 1847168,
105
+ "step": 24
106
+ },
107
+ {
108
+ "epoch": 0.1038442336495257,
109
+ "grad_norm": 0.028687166050076485,
110
+ "learning_rate": 4.867745762486861e-05,
111
+ "loss": 2.4515,
112
+ "num_input_tokens_seen": 2002624,
113
+ "step": 26
114
+ },
115
+ {
116
+ "epoch": 0.11183225162256615,
117
+ "grad_norm": 0.02790650725364685,
118
+ "learning_rate": 4.846834644134686e-05,
119
+ "loss": 2.4309,
120
+ "num_input_tokens_seen": 2162992,
121
+ "step": 28
122
+ },
123
+ {
124
+ "epoch": 0.11982026959560658,
125
+ "grad_norm": 0.0356418751180172,
126
+ "learning_rate": 4.8244412147206284e-05,
127
+ "loss": 2.4502,
128
+ "num_input_tokens_seen": 2304544,
129
+ "step": 30
130
+ },
131
+ {
132
+ "epoch": 0.12780828756864704,
133
+ "grad_norm": 0.03653424605727196,
134
+ "learning_rate": 4.800579618414676e-05,
135
+ "loss": 2.4509,
136
+ "num_input_tokens_seen": 2444336,
137
+ "step": 32
138
+ },
139
+ {
140
+ "epoch": 0.13579630554168748,
141
+ "grad_norm": 0.03637837618589401,
142
+ "learning_rate": 4.775264926712489e-05,
143
+ "loss": 2.4778,
144
+ "num_input_tokens_seen": 2598848,
145
+ "step": 34
146
+ },
147
+ {
148
+ "epoch": 0.14378432351472792,
149
+ "grad_norm": 0.03624645993113518,
150
+ "learning_rate": 4.7485131289159276e-05,
151
+ "loss": 2.3898,
152
+ "num_input_tokens_seen": 2745520,
153
+ "step": 36
154
+ },
155
+ {
156
+ "epoch": 0.15177234148776836,
157
+ "grad_norm": 0.039211999624967575,
158
+ "learning_rate": 4.720341122033862e-05,
159
+ "loss": 2.5515,
160
+ "num_input_tokens_seen": 2913392,
161
+ "step": 38
162
+ },
163
+ {
164
+ "epoch": 0.1597603594608088,
165
+ "grad_norm": 0.04233478382229805,
166
+ "learning_rate": 4.690766700109659e-05,
167
+ "loss": 2.4457,
168
+ "num_input_tokens_seen": 3053600,
169
+ "step": 40
170
+ },
171
+ {
172
+ "epoch": 0.16774837743384924,
173
+ "grad_norm": 0.03884820267558098,
174
+ "learning_rate": 4.659808542982088e-05,
175
+ "loss": 2.4336,
176
+ "num_input_tokens_seen": 3199248,
177
+ "step": 42
178
+ },
179
+ {
180
+ "epoch": 0.17573639540688968,
181
+ "grad_norm": 0.04048454761505127,
182
+ "learning_rate": 4.6274862044867304e-05,
183
+ "loss": 2.4655,
184
+ "num_input_tokens_seen": 3324464,
185
+ "step": 44
186
+ },
187
+ {
188
+ "epoch": 0.18372441337993012,
189
+ "grad_norm": 0.034897685050964355,
190
+ "learning_rate": 4.593820100105355e-05,
191
+ "loss": 2.4566,
192
+ "num_input_tokens_seen": 3485088,
193
+ "step": 46
194
+ },
195
+ {
196
+ "epoch": 0.19171243135297053,
197
+ "grad_norm": 0.04081996530294418,
198
+ "learning_rate": 4.558831494071069e-05,
199
+ "loss": 2.4912,
200
+ "num_input_tokens_seen": 3635152,
201
+ "step": 48
202
+ },
203
+ {
204
+ "epoch": 0.19970044932601097,
205
+ "grad_norm": 0.03764244541525841,
206
+ "learning_rate": 4.522542485937369e-05,
207
+ "loss": 2.4185,
208
+ "num_input_tokens_seen": 3787152,
209
+ "step": 50
210
+ },
211
+ {
212
+ "epoch": 0.2076884672990514,
213
+ "grad_norm": 0.04559960216283798,
214
+ "learning_rate": 4.484975996619589e-05,
215
+ "loss": 2.4221,
216
+ "num_input_tokens_seen": 3931280,
217
+ "step": 52
218
+ },
219
+ {
220
+ "epoch": 0.21567648527209185,
221
+ "grad_norm": 0.03996207192540169,
222
+ "learning_rate": 4.4461557539175594e-05,
223
+ "loss": 2.4091,
224
+ "num_input_tokens_seen": 4083536,
225
+ "step": 54
226
+ },
227
+ {
228
+ "epoch": 0.2236645032451323,
229
+ "grad_norm": 0.042057909071445465,
230
+ "learning_rate": 4.40610627752862e-05,
231
+ "loss": 2.4697,
232
+ "num_input_tokens_seen": 4239440,
233
+ "step": 56
234
+ },
235
+ {
236
+ "epoch": 0.23165252121817273,
237
+ "grad_norm": 0.04598196595907211,
238
+ "learning_rate": 4.3648528635604556e-05,
239
+ "loss": 2.3675,
240
+ "num_input_tokens_seen": 4369008,
241
+ "step": 58
242
+ },
243
+ {
244
+ "epoch": 0.23964053919121317,
245
+ "grad_norm": 0.03921639174222946,
246
+ "learning_rate": 4.3224215685535294e-05,
247
+ "loss": 2.4421,
248
+ "num_input_tokens_seen": 4525680,
249
+ "step": 60
250
+ },
251
+ {
252
+ "epoch": 0.2476285571642536,
253
+ "grad_norm": 0.03894852474331856,
254
+ "learning_rate": 4.278839193023214e-05,
255
+ "loss": 2.3973,
256
+ "num_input_tokens_seen": 4683360,
257
+ "step": 62
258
+ },
259
+ {
260
+ "epoch": 0.2556165751372941,
261
+ "grad_norm": 0.048196952790021896,
262
+ "learning_rate": 4.234133264532012e-05,
263
+ "loss": 2.5022,
264
+ "num_input_tokens_seen": 4833520,
265
+ "step": 64
266
+ },
267
+ {
268
+ "epoch": 0.2636045931103345,
269
+ "grad_norm": 0.03995994105935097,
270
+ "learning_rate": 4.188332020302561e-05,
271
+ "loss": 2.3089,
272
+ "num_input_tokens_seen": 4974480,
273
+ "step": 66
274
+ },
275
+ {
276
+ "epoch": 0.27159261108337496,
277
+ "grad_norm": 0.047357227653265,
278
+ "learning_rate": 4.1414643893823914e-05,
279
+ "loss": 2.3378,
280
+ "num_input_tokens_seen": 5099616,
281
+ "step": 68
282
+ },
283
+ {
284
+ "epoch": 0.2795806290564154,
285
+ "grad_norm": 0.046735942363739014,
286
+ "learning_rate": 4.093559974371725e-05,
287
+ "loss": 2.4277,
288
+ "num_input_tokens_seen": 5244416,
289
+ "step": 70
290
+ },
291
+ {
292
+ "epoch": 0.28756864702945584,
293
+ "grad_norm": 0.04764910414814949,
294
+ "learning_rate": 4.044649032725836e-05,
295
+ "loss": 2.3468,
296
+ "num_input_tokens_seen": 5392128,
297
+ "step": 72
298
+ },
299
+ {
300
+ "epoch": 0.2955566650024963,
301
+ "grad_norm": 0.04130158573389053,
302
+ "learning_rate": 3.9947624576437975e-05,
303
+ "loss": 2.3907,
304
+ "num_input_tokens_seen": 5545216,
305
+ "step": 74
306
+ },
307
+ {
308
+ "epoch": 0.3035446829755367,
309
+ "grad_norm": 0.048350926488637924,
310
+ "learning_rate": 3.943931758555669e-05,
311
+ "loss": 2.3878,
312
+ "num_input_tokens_seen": 5712080,
313
+ "step": 76
314
+ },
315
+ {
316
+ "epoch": 0.31153270094857716,
317
+ "grad_norm": 0.052576858550310135,
318
+ "learning_rate": 3.8921890412204705e-05,
319
+ "loss": 2.3853,
320
+ "num_input_tokens_seen": 5858800,
321
+ "step": 78
322
+ },
323
+ {
324
+ "epoch": 0.3195207189216176,
325
+ "grad_norm": 0.04666517302393913,
326
+ "learning_rate": 3.8395669874474915e-05,
327
+ "loss": 2.3765,
328
+ "num_input_tokens_seen": 6008784,
329
+ "step": 80
330
+ },
331
+ {
332
+ "epoch": 0.32750873689465804,
333
+ "grad_norm": 0.045086730271577835,
334
+ "learning_rate": 3.786098834453766e-05,
335
+ "loss": 2.4277,
336
+ "num_input_tokens_seen": 6155536,
337
+ "step": 82
338
+ },
339
+ {
340
+ "epoch": 0.3354967548676985,
341
+ "grad_norm": 0.0495474711060524,
342
+ "learning_rate": 3.731818353870729e-05,
343
+ "loss": 2.3935,
344
+ "num_input_tokens_seen": 6306576,
345
+ "step": 84
346
+ },
347
+ {
348
+ "epoch": 0.3434847728407389,
349
+ "grad_norm": 0.0486772395670414,
350
+ "learning_rate": 3.6767598304133324e-05,
351
+ "loss": 2.4679,
352
+ "num_input_tokens_seen": 6470448,
353
+ "step": 86
354
+ },
355
+ {
356
+ "epoch": 0.35147279081377936,
357
+ "grad_norm": 0.05734413489699364,
358
+ "learning_rate": 3.6209580402250815e-05,
359
+ "loss": 2.3868,
360
+ "num_input_tokens_seen": 6616864,
361
+ "step": 88
362
+ },
363
+ {
364
+ "epoch": 0.3594608087868198,
365
+ "grad_norm": 0.04023748263716698,
366
+ "learning_rate": 3.564448228912682e-05,
367
+ "loss": 2.3859,
368
+ "num_input_tokens_seen": 6786448,
369
+ "step": 90
370
+ },
371
+ {
372
+ "epoch": 0.36744882675986024,
373
+ "grad_norm": 0.04801137000322342,
374
+ "learning_rate": 3.507266089284157e-05,
375
+ "loss": 2.4216,
376
+ "num_input_tokens_seen": 6954720,
377
+ "step": 92
378
+ },
379
+ {
380
+ "epoch": 0.3754368447329007,
381
+ "grad_norm": 0.05846545100212097,
382
+ "learning_rate": 3.4494477388045035e-05,
383
+ "loss": 2.4611,
384
+ "num_input_tokens_seen": 7094208,
385
+ "step": 94
386
+ },
387
+ {
388
+ "epoch": 0.38342486270594106,
389
+ "grad_norm": 0.04973718896508217,
390
+ "learning_rate": 3.3910296967831266e-05,
391
+ "loss": 2.3384,
392
+ "num_input_tokens_seen": 7227440,
393
+ "step": 96
394
+ },
395
+ {
396
+ "epoch": 0.3914128806789815,
397
+ "grad_norm": 0.0684569925069809,
398
+ "learning_rate": 3.332048861307467e-05,
399
+ "loss": 2.4074,
400
+ "num_input_tokens_seen": 7362608,
401
+ "step": 98
402
+ },
403
+ {
404
+ "epoch": 0.39940089865202194,
405
+ "grad_norm": 0.05771000683307648,
406
+ "learning_rate": 3.272542485937369e-05,
407
+ "loss": 2.3572,
408
+ "num_input_tokens_seen": 7515024,
409
+ "step": 100
410
+ },
411
+ {
412
+ "epoch": 0.4073889166250624,
413
+ "grad_norm": 0.043234046548604965,
414
+ "learning_rate": 3.21254815617494e-05,
415
+ "loss": 2.3579,
416
+ "num_input_tokens_seen": 7670624,
417
+ "step": 102
418
+ },
419
+ {
420
+ "epoch": 0.4153769345981028,
421
+ "grad_norm": 0.06148277968168259,
422
+ "learning_rate": 3.152103765724743e-05,
423
+ "loss": 2.3728,
424
+ "num_input_tokens_seen": 7817648,
425
+ "step": 104
426
+ },
427
+ {
428
+ "epoch": 0.42336495257114326,
429
+ "grad_norm": 0.04587692394852638,
430
+ "learning_rate": 3.091247492559312e-05,
431
+ "loss": 2.2712,
432
+ "num_input_tokens_seen": 7968432,
433
+ "step": 106
434
+ },
435
+ {
436
+ "epoch": 0.4313529705441837,
437
+ "grad_norm": 0.05067060887813568,
438
+ "learning_rate": 3.0300177748051373e-05,
439
+ "loss": 2.4054,
440
+ "num_input_tokens_seen": 8127312,
441
+ "step": 108
442
+ },
443
+ {
444
+ "epoch": 0.43934098851722414,
445
+ "grad_norm": 0.04505151882767677,
446
+ "learning_rate": 2.9684532864643122e-05,
447
+ "loss": 2.3966,
448
+ "num_input_tokens_seen": 8275008,
449
+ "step": 110
450
+ },
451
+ {
452
+ "epoch": 0.4473290064902646,
453
+ "grad_norm": 0.058195654302835464,
454
+ "learning_rate": 2.9065929129872094e-05,
455
+ "loss": 2.4478,
456
+ "num_input_tokens_seen": 8425536,
457
+ "step": 112
458
+ },
459
+ {
460
+ "epoch": 0.455317024463305,
461
+ "grad_norm": 0.05548926442861557,
462
+ "learning_rate": 2.844475726711595e-05,
463
+ "loss": 2.4386,
464
+ "num_input_tokens_seen": 8573776,
465
+ "step": 114
466
+ },
467
+ {
468
+ "epoch": 0.46330504243634546,
469
+ "grad_norm": 0.05548759549856186,
470
+ "learning_rate": 2.782140962183704e-05,
471
+ "loss": 2.4176,
472
+ "num_input_tokens_seen": 8729200,
473
+ "step": 116
474
+ },
475
+ {
476
+ "epoch": 0.4712930604093859,
477
+ "grad_norm": 0.05361337587237358,
478
+ "learning_rate": 2.7196279913768584e-05,
479
+ "loss": 2.3505,
480
+ "num_input_tokens_seen": 8884720,
481
+ "step": 118
482
+ },
483
+ {
484
+ "epoch": 0.47928107838242634,
485
+ "grad_norm": 0.05254572257399559,
486
+ "learning_rate": 2.656976298823284e-05,
487
+ "loss": 2.3867,
488
+ "num_input_tokens_seen": 9035088,
489
+ "step": 120
490
+ },
491
+ {
492
+ "epoch": 0.4872690963554668,
493
+ "grad_norm": 0.051810409873723984,
494
+ "learning_rate": 2.594225456674837e-05,
495
+ "loss": 2.4321,
496
+ "num_input_tokens_seen": 9188528,
497
+ "step": 122
498
+ },
499
+ {
500
+ "epoch": 0.4952571143285072,
501
+ "grad_norm": 0.05497866868972778,
502
+ "learning_rate": 2.531415099708382e-05,
503
+ "loss": 2.4392,
504
+ "num_input_tokens_seen": 9352752,
505
+ "step": 124
506
+ },
507
+ {
508
+ "epoch": 0.5032451323015477,
509
+ "grad_norm": 0.06646806746721268,
510
+ "learning_rate": 2.4685849002916183e-05,
511
+ "loss": 2.4005,
512
+ "num_input_tokens_seen": 9497840,
513
+ "step": 126
514
+ },
515
+ {
516
+ "epoch": 0.5112331502745882,
517
+ "grad_norm": 0.05319731682538986,
518
+ "learning_rate": 2.4057745433251635e-05,
519
+ "loss": 2.309,
520
+ "num_input_tokens_seen": 9654816,
521
+ "step": 128
522
+ },
523
+ {
524
+ "epoch": 0.5192211682476285,
525
+ "grad_norm": 0.07142467051744461,
526
+ "learning_rate": 2.3430237011767167e-05,
527
+ "loss": 2.522,
528
+ "num_input_tokens_seen": 9814944,
529
+ "step": 130
530
+ },
531
+ {
532
+ "epoch": 0.527209186220669,
533
+ "grad_norm": 0.06771399825811386,
534
+ "learning_rate": 2.280372008623142e-05,
535
+ "loss": 2.4151,
536
+ "num_input_tokens_seen": 9951952,
537
+ "step": 132
538
+ },
539
+ {
540
+ "epoch": 0.5351972041937094,
541
+ "grad_norm": 0.062045566737651825,
542
+ "learning_rate": 2.217859037816296e-05,
543
+ "loss": 2.3723,
544
+ "num_input_tokens_seen": 10112272,
545
+ "step": 134
546
+ },
547
+ {
548
+ "epoch": 0.5431852221667499,
549
+ "grad_norm": 0.05891815572977066,
550
+ "learning_rate": 2.155524273288405e-05,
551
+ "loss": 2.416,
552
+ "num_input_tokens_seen": 10281344,
553
+ "step": 136
554
+ },
555
+ {
556
+ "epoch": 0.5511732401397903,
557
+ "grad_norm": 0.06526540219783783,
558
+ "learning_rate": 2.0934070870127912e-05,
559
+ "loss": 2.5335,
560
+ "num_input_tokens_seen": 10409584,
561
+ "step": 138
562
+ },
563
+ {
564
+ "epoch": 0.5591612581128308,
565
+ "grad_norm": 0.0685892105102539,
566
+ "learning_rate": 2.031546713535688e-05,
567
+ "loss": 2.4946,
568
+ "num_input_tokens_seen": 10573840,
569
+ "step": 140
570
+ },
571
+ {
572
+ "epoch": 0.5671492760858712,
573
+ "grad_norm": 0.05832570418715477,
574
+ "learning_rate": 1.969982225194864e-05,
575
+ "loss": 2.3169,
576
+ "num_input_tokens_seen": 10729760,
577
+ "step": 142
578
+ },
579
+ {
580
+ "epoch": 0.5751372940589117,
581
+ "grad_norm": 0.05707252770662308,
582
+ "learning_rate": 1.908752507440689e-05,
583
+ "loss": 2.3797,
584
+ "num_input_tokens_seen": 10881664,
585
+ "step": 144
586
+ },
587
+ {
588
+ "epoch": 0.5831253120319521,
589
+ "grad_norm": 0.05318214371800423,
590
+ "learning_rate": 1.8478962342752583e-05,
591
+ "loss": 2.4498,
592
+ "num_input_tokens_seen": 11034352,
593
+ "step": 146
594
+ },
595
+ {
596
+ "epoch": 0.5911133300049926,
597
+ "grad_norm": 0.06147105619311333,
598
+ "learning_rate": 1.7874518438250597e-05,
599
+ "loss": 2.3818,
600
+ "num_input_tokens_seen": 11175952,
601
+ "step": 148
602
+ },
603
+ {
604
+ "epoch": 0.5991013479780329,
605
+ "grad_norm": 0.05764273181557655,
606
+ "learning_rate": 1.7274575140626318e-05,
607
+ "loss": 2.3818,
608
+ "num_input_tokens_seen": 11314112,
609
+ "step": 150
610
+ },
611
+ {
612
+ "epoch": 0.6070893659510734,
613
+ "grad_norm": 0.05358808860182762,
614
+ "learning_rate": 1.6679511386925337e-05,
615
+ "loss": 2.4073,
616
+ "num_input_tokens_seen": 11481792,
617
+ "step": 152
618
+ },
619
+ {
620
+ "epoch": 0.6150773839241138,
621
+ "grad_norm": 0.05264300853013992,
622
+ "learning_rate": 1.6089703032168733e-05,
623
+ "loss": 2.3467,
624
+ "num_input_tokens_seen": 11649040,
625
+ "step": 154
626
+ },
627
+ {
628
+ "epoch": 0.6230654018971543,
629
+ "grad_norm": 0.060223497450351715,
630
+ "learning_rate": 1.5505522611954975e-05,
631
+ "loss": 2.4047,
632
+ "num_input_tokens_seen": 11791488,
633
+ "step": 156
634
+ },
635
+ {
636
+ "epoch": 0.6310534198701947,
637
+ "grad_norm": 0.0639820247888565,
638
+ "learning_rate": 1.4927339107158437e-05,
639
+ "loss": 2.2998,
640
+ "num_input_tokens_seen": 11940144,
641
+ "step": 158
642
+ },
643
+ {
644
+ "epoch": 0.6390414378432352,
645
+ "grad_norm": 0.053195178508758545,
646
+ "learning_rate": 1.4355517710873184e-05,
647
+ "loss": 2.383,
648
+ "num_input_tokens_seen": 12097648,
649
+ "step": 160
650
+ },
651
+ {
652
+ "epoch": 0.6470294558162756,
653
+ "grad_norm": 0.059055812656879425,
654
+ "learning_rate": 1.3790419597749199e-05,
655
+ "loss": 2.3525,
656
+ "num_input_tokens_seen": 12253152,
657
+ "step": 162
658
+ },
659
+ {
660
+ "epoch": 0.6550174737893161,
661
+ "grad_norm": 0.07648273557424545,
662
+ "learning_rate": 1.3232401695866687e-05,
663
+ "loss": 2.4133,
664
+ "num_input_tokens_seen": 12410080,
665
+ "step": 164
666
+ },
667
+ {
668
+ "epoch": 0.6630054917623565,
669
+ "grad_norm": 0.06658945977687836,
670
+ "learning_rate": 1.2681816461292715e-05,
671
+ "loss": 2.3677,
672
+ "num_input_tokens_seen": 12544848,
673
+ "step": 166
674
+ },
675
+ {
676
+ "epoch": 0.670993509735397,
677
+ "grad_norm": 0.06263954192399979,
678
+ "learning_rate": 1.2139011655462337e-05,
679
+ "loss": 2.4215,
680
+ "num_input_tokens_seen": 12701808,
681
+ "step": 168
682
+ },
683
+ {
684
+ "epoch": 0.6789815277084373,
685
+ "grad_norm": 0.0712100937962532,
686
+ "learning_rate": 1.1604330125525079e-05,
687
+ "loss": 2.3162,
688
+ "num_input_tokens_seen": 12845840,
689
+ "step": 170
690
+ },
691
+ {
692
+ "epoch": 0.6869695456814778,
693
+ "grad_norm": 0.06635987758636475,
694
+ "learning_rate": 1.107810958779531e-05,
695
+ "loss": 2.4129,
696
+ "num_input_tokens_seen": 12988144,
697
+ "step": 172
698
+ },
699
+ {
700
+ "epoch": 0.6949575636545182,
701
+ "grad_norm": 0.07223087549209595,
702
+ "learning_rate": 1.0560682414443315e-05,
703
+ "loss": 2.4046,
704
+ "num_input_tokens_seen": 13129808,
705
+ "step": 174
706
+ },
707
+ {
708
+ "epoch": 0.7029455816275587,
709
+ "grad_norm": 0.08978503942489624,
710
+ "learning_rate": 1.0052375423562038e-05,
711
+ "loss": 2.3596,
712
+ "num_input_tokens_seen": 13269264,
713
+ "step": 176
714
+ },
715
+ {
716
+ "epoch": 0.7109335996005991,
717
+ "grad_norm": 0.06082882732152939,
718
+ "learning_rate": 9.553509672741645e-06,
719
+ "loss": 2.3047,
720
+ "num_input_tokens_seen": 13417184,
721
+ "step": 178
722
+ },
723
+ {
724
+ "epoch": 0.7189216175736396,
725
+ "grad_norm": 0.07053129374980927,
726
+ "learning_rate": 9.064400256282757e-06,
727
+ "loss": 2.4177,
728
+ "num_input_tokens_seen": 13561008,
729
+ "step": 180
730
+ },
731
+ {
732
+ "epoch": 0.72690963554668,
733
+ "grad_norm": 0.0724414512515068,
734
+ "learning_rate": 8.585356106176094e-06,
735
+ "loss": 2.4451,
736
+ "num_input_tokens_seen": 13706976,
737
+ "step": 182
738
+ },
739
+ {
740
+ "epoch": 0.7348976535197205,
741
+ "grad_norm": 0.07262030243873596,
742
+ "learning_rate": 8.116679796974388e-06,
743
+ "loss": 2.3997,
744
+ "num_input_tokens_seen": 13835936,
745
+ "step": 184
746
+ },
747
+ {
748
+ "epoch": 0.7428856714927609,
749
+ "grad_norm": 0.07069452106952667,
750
+ "learning_rate": 7.65866735467988e-06,
751
+ "loss": 2.412,
752
+ "num_input_tokens_seen": 13962144,
753
+ "step": 186
754
+ },
755
+ {
756
+ "epoch": 0.7508736894658014,
757
+ "grad_norm": 0.053225237876176834,
758
+ "learning_rate": 7.211608069767867e-06,
759
+ "loss": 2.4054,
760
+ "num_input_tokens_seen": 14128800,
761
+ "step": 188
762
+ },
763
+ {
764
+ "epoch": 0.7588617074388417,
765
+ "grad_norm": 0.05883209779858589,
766
+ "learning_rate": 6.775784314464717e-06,
767
+ "loss": 2.4417,
768
+ "num_input_tokens_seen": 14278288,
769
+ "step": 190
770
+ },
771
+ {
772
+ "epoch": 0.7668497254118821,
773
+ "grad_norm": 0.0630330890417099,
774
+ "learning_rate": 6.3514713643954475e-06,
775
+ "loss": 2.3358,
776
+ "num_input_tokens_seen": 14429936,
777
+ "step": 192
778
+ },
779
+ {
780
+ "epoch": 0.7748377433849226,
781
+ "grad_norm": 0.06319437175989151,
782
+ "learning_rate": 5.9389372247138e-06,
783
+ "loss": 2.5089,
784
+ "num_input_tokens_seen": 14567552,
785
+ "step": 194
786
+ },
787
+ {
788
+ "epoch": 0.782825761357963,
789
+ "grad_norm": 0.0682671070098877,
790
+ "learning_rate": 5.538442460824417e-06,
791
+ "loss": 2.4246,
792
+ "num_input_tokens_seen": 14704528,
793
+ "step": 196
794
+ },
795
+ {
796
+ "epoch": 0.7908137793310035,
797
+ "grad_norm": 0.06514116376638412,
798
+ "learning_rate": 5.150240033804116e-06,
799
+ "loss": 2.369,
800
+ "num_input_tokens_seen": 14846192,
801
+ "step": 198
802
+ },
803
+ {
804
+ "epoch": 0.7988017973040439,
805
+ "grad_norm": 0.05886775627732277,
806
+ "learning_rate": 4.7745751406263165e-06,
807
+ "loss": 2.4204,
808
+ "num_input_tokens_seen": 14993936,
809
+ "step": 200
810
+ },
811
+ {
812
+ "epoch": 0.8067898152770844,
813
+ "grad_norm": 0.06780359148979187,
814
+ "learning_rate": 4.411685059289314e-06,
815
+ "loss": 2.3141,
816
+ "num_input_tokens_seen": 15136240,
817
+ "step": 202
818
+ },
819
+ {
820
+ "epoch": 0.8147778332501248,
821
+ "grad_norm": 0.06595364212989807,
822
+ "learning_rate": 4.061798998946459e-06,
823
+ "loss": 2.3841,
824
+ "num_input_tokens_seen": 15274672,
825
+ "step": 204
826
+ },
827
+ {
828
+ "epoch": 0.8227658512231653,
829
+ "grad_norm": 0.06310597062110901,
830
+ "learning_rate": 3.725137955132707e-06,
831
+ "loss": 2.3653,
832
+ "num_input_tokens_seen": 15417920,
833
+ "step": 206
834
+ },
835
+ {
836
+ "epoch": 0.8307538691962056,
837
+ "grad_norm": 0.06177780404686928,
838
+ "learning_rate": 3.4019145701791184e-06,
839
+ "loss": 2.4382,
840
+ "num_input_tokens_seen": 15557920,
841
+ "step": 208
842
+ },
843
+ {
844
+ "epoch": 0.8387418871692461,
845
+ "grad_norm": 0.0720074325799942,
846
+ "learning_rate": 3.092332998903416e-06,
847
+ "loss": 2.4455,
848
+ "num_input_tokens_seen": 15687968,
849
+ "step": 210
850
+ },
851
+ {
852
+ "epoch": 0.8467299051422865,
853
+ "grad_norm": 0.05820966139435768,
854
+ "learning_rate": 2.7965887796613884e-06,
855
+ "loss": 2.4322,
856
+ "num_input_tokens_seen": 15843072,
857
+ "step": 212
858
+ },
859
+ {
860
+ "epoch": 0.854717923115327,
861
+ "grad_norm": 0.07062980532646179,
862
+ "learning_rate": 2.514868710840723e-06,
863
+ "loss": 2.4433,
864
+ "num_input_tokens_seen": 16002112,
865
+ "step": 214
866
+ },
867
+ {
868
+ "epoch": 0.8627059410883674,
869
+ "grad_norm": 0.07145073264837265,
870
+ "learning_rate": 2.2473507328751086e-06,
871
+ "loss": 2.4399,
872
+ "num_input_tokens_seen": 16135328,
873
+ "step": 216
874
+ },
875
+ {
876
+ "epoch": 0.8706939590614079,
877
+ "grad_norm": 0.06476866453886032,
878
+ "learning_rate": 1.9942038158532407e-06,
879
+ "loss": 2.378,
880
+ "num_input_tokens_seen": 16293008,
881
+ "step": 218
882
+ },
883
+ {
884
+ "epoch": 0.8786819770344483,
885
+ "grad_norm": 0.05755852535367012,
886
+ "learning_rate": 1.7555878527937164e-06,
887
+ "loss": 2.3874,
888
+ "num_input_tokens_seen": 16441024,
889
+ "step": 220
890
+ },
891
+ {
892
+ "epoch": 0.8866699950074888,
893
+ "grad_norm": 0.059895843267440796,
894
+ "learning_rate": 1.5316535586531483e-06,
895
+ "loss": 2.4028,
896
+ "num_input_tokens_seen": 16587616,
897
+ "step": 222
898
+ },
899
+ {
900
+ "epoch": 0.8946580129805292,
901
+ "grad_norm": 0.055422358214855194,
902
+ "learning_rate": 1.3225423751313942e-06,
903
+ "loss": 2.3372,
904
+ "num_input_tokens_seen": 16730800,
905
+ "step": 224
906
+ },
907
+ {
908
+ "epoch": 0.9026460309535697,
909
+ "grad_norm": 0.06502599269151688,
910
+ "learning_rate": 1.1283863813339263e-06,
911
+ "loss": 2.4288,
912
+ "num_input_tokens_seen": 16886144,
913
+ "step": 226
914
+ },
915
+ {
916
+ "epoch": 0.91063404892661,
917
+ "grad_norm": 0.06254018843173981,
918
+ "learning_rate": 9.493082103478517e-07,
919
+ "loss": 2.3483,
920
+ "num_input_tokens_seen": 17038816,
921
+ "step": 228
922
+ },
923
+ {
924
+ "epoch": 0.9186220668996505,
925
+ "grad_norm": 0.08103015273809433,
926
+ "learning_rate": 7.854209717842231e-07,
927
+ "loss": 2.4648,
928
+ "num_input_tokens_seen": 17171792,
929
+ "step": 230
930
+ },
931
+ {
932
+ "epoch": 0.9266100848726909,
933
+ "grad_norm": 0.06423594057559967,
934
+ "learning_rate": 6.368281803355691e-07,
935
+ "loss": 2.4327,
936
+ "num_input_tokens_seen": 17324752,
937
+ "step": 232
938
+ },
939
+ {
940
+ "epoch": 0.9345981028457314,
941
+ "grad_norm": 0.06621188670396805,
942
+ "learning_rate": 5.036236903938285e-07,
943
+ "loss": 2.3037,
944
+ "num_input_tokens_seen": 17485168,
945
+ "step": 234
946
+ },
947
+ {
948
+ "epoch": 0.9425861208187718,
949
+ "grad_norm": 0.06229124590754509,
950
+ "learning_rate": 3.8589163676986674e-07,
951
+ "loss": 2.4094,
952
+ "num_input_tokens_seen": 17632832,
953
+ "step": 236
954
+ },
955
+ {
956
+ "epoch": 0.9505741387918123,
957
+ "grad_norm": 0.07828541845083237,
958
+ "learning_rate": 2.8370638155215123e-07,
959
+ "loss": 2.3348,
960
+ "num_input_tokens_seen": 17781744,
961
+ "step": 238
962
+ },
963
+ {
964
+ "epoch": 0.9585621567648527,
965
+ "grad_norm": 0.049717940390110016,
966
+ "learning_rate": 1.9713246713805588e-07,
967
+ "loss": 2.3979,
968
+ "num_input_tokens_seen": 17959392,
969
+ "step": 240
970
+ },
971
+ {
972
+ "epoch": 0.9665501747378932,
973
+ "grad_norm": 0.06252018362283707,
974
+ "learning_rate": 1.2622457546749567e-07,
975
+ "loss": 2.346,
976
+ "num_input_tokens_seen": 18121744,
977
+ "step": 242
978
+ },
979
+ {
980
+ "epoch": 0.9745381927109336,
981
+ "grad_norm": 0.06641443818807602,
982
+ "learning_rate": 7.102749348465165e-08,
983
+ "loss": 2.4486,
984
+ "num_input_tokens_seen": 18273856,
985
+ "step": 244
986
+ },
987
+ {
988
+ "epoch": 0.982526210683974,
989
+ "grad_norm": 0.052787262946367264,
990
+ "learning_rate": 3.157608484956332e-08,
991
+ "loss": 2.3959,
992
+ "num_input_tokens_seen": 18407456,
993
+ "step": 246
994
+ },
995
+ {
996
+ "epoch": 0.9905142286570144,
997
+ "grad_norm": 0.058858100324869156,
998
+ "learning_rate": 7.895267917501504e-09,
999
+ "loss": 2.4287,
1000
+ "num_input_tokens_seen": 18555776,
1001
+ "step": 248
1002
+ },
1003
+ {
1004
+ "epoch": 0.9985022466300549,
1005
+ "grad_norm": 0.07202674448490143,
1006
+ "learning_rate": 0.0,
1007
+ "loss": 2.4911,
1008
+ "num_input_tokens_seen": 18683616,
1009
+ "step": 250
1010
+ },
1011
+ {
1012
+ "epoch": 0.9985022466300549,
1013
+ "num_input_tokens_seen": 18683616,
1014
+ "step": 250,
1015
+ "total_flos": 1.599323192678744e+18,
1016
+ "train_loss": 2.4129163398742675,
1017
+ "train_runtime": 18708.0864,
1018
+ "train_samples_per_second": 0.214,
1019
+ "train_steps_per_second": 0.013
1020
+ }
1021
+ ],
1022
+ "logging_steps": 2,
1023
+ "max_steps": 250,
1024
+ "num_input_tokens_seen": 18683616,
1025
+ "num_train_epochs": 1,
1026
+ "save_steps": 1000,
1027
+ "stateful_callbacks": {
1028
+ "TrainerControl": {
1029
+ "args": {
1030
+ "should_epoch_stop": false,
1031
+ "should_evaluate": false,
1032
+ "should_log": false,
1033
+ "should_save": true,
1034
+ "should_training_stop": true
1035
+ },
1036
+ "attributes": {}
1037
+ }
1038
+ },
1039
+ "total_flos": 1.599323192678744e+18,
1040
+ "train_batch_size": 2,
1041
+ "trial_name": null,
1042
+ "trial_params": null
1043
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d83dfded8f6f28abb3b96317b9a8d68c218c87cbec77be722fe4ea54ebd69b6d
3
+ size 5432
training_loss.png ADDED
vocab.json ADDED
The diff for this file is too large to render. See raw diff