ben81828 commited on
Commit
f742815
·
verified ·
1 Parent(s): ca23548

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ec17bc88ac76153637df1b04176014909e961cdf177b910eb472c5e048439bb
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd961b0e78417f1e80902a52b4b9a133d9724bddd4f50db6ac185504a1aebd52
3
  size 18516456
last-checkpoint/global_step100/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1b5222ece9c07bc0707840fd209d23ed3425181c60bbe9066a81da1cc913f8a
3
+ size 27700976
last-checkpoint/global_step100/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf6bb8fd7d7e600dab44cd221b6f39d1fc98b2b65a22c8ff0c03f1f91b855c07
3
+ size 27700976
last-checkpoint/global_step100/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d55864a517cc73f6af60442807993dfcab4932a335a2dc5a1ce07e58fc3fedbc
3
+ size 27700976
last-checkpoint/global_step100/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c6617a438e97dca9ca57f429bb94e72f3270e48616e5ddabd3c45eef92888c
3
+ size 27700976
last-checkpoint/global_step100/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77ab030b3fb16c943c20557696294e261908aa49808993525db2314febcd364c
3
+ size 411571
last-checkpoint/global_step100/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98a12a65975be5132e3b8312e6c4ad5570bb2e62a73e91d7be7eb4d7809eb237
3
+ size 411507
last-checkpoint/global_step100/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf6a973959fae837554064cd65731d8cf436b2cf978d2efbc00bf7ce6b54eacb
3
+ size 411507
last-checkpoint/global_step100/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d582bb0bfd605fc477447d0f280945126b1ed22748ebad3a0689426e5c97f8c5
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step50
 
1
+ global_step100
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbe0d720c4c75a6a04213fa3b64bacbe794718a53e2b56ebb67a1a795014dfad
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:308f94f9a5c24e1bad5c393d56ae7af7782600f4e791d9c6ac35b22fff2105b6
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72452d3138d0ca2ff89429e3294a834ae7a68e8596fc757735ca56ae52509d57
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b056f3c23cb32dc77a2ec9e7651e0b64e4440e21f0fdf969b86bfc56a1cbdf06
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f36e306fb8ebcf53a167bfd6c9af74db410a269ada1e619e3e816f5269543b9d
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3f8a05714bc528f4885a2816181652f2303b3e8150f89b56aaee6bec56aa520
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bb47ce0c6f815a6f8302b0e3819b4c2315ca71dae3138d97fdceb765cdd0a039
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f755bd3c330281961e5c03af9d10ce8c1e1678619d384f6f1fd5fd7dce2ff50
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ab82860e8f4753962498aa91f545ddf5d94711cc852b62536116736b2908eac
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e451a9e086b06d7c667be8442b2115f5c088953bade0b625e61f2ce5c7fd404
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.9191630482673645,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-50",
4
- "epoch": 0.025753283543651816,
5
  "eval_steps": 50,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -96,11 +96,100 @@
96
  "eval_steps_per_second": 0.41,
97
  "num_input_tokens_seen": 584856,
98
  "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
99
  }
100
  ],
101
  "logging_steps": 5,
102
  "max_steps": 3400,
103
- "num_input_tokens_seen": 584856,
104
  "num_train_epochs": 2,
105
  "save_steps": 50,
106
  "stateful_callbacks": {
@@ -115,7 +204,7 @@
115
  "attributes": {}
116
  }
117
  },
118
- "total_flos": 32796859170816.0,
119
  "train_batch_size": 1,
120
  "trial_name": null,
121
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.9077914953231812,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-100",
4
+ "epoch": 0.05150656708730363,
5
  "eval_steps": 50,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
96
  "eval_steps_per_second": 0.41,
97
  "num_input_tokens_seen": 584856,
98
  "step": 50
99
+ },
100
+ {
101
+ "epoch": 0.028328611898016998,
102
+ "grad_norm": 0.8294990584587586,
103
+ "learning_rate": 3.235294117647059e-05,
104
+ "loss": 0.9009,
105
+ "num_input_tokens_seen": 643344,
106
+ "step": 55
107
+ },
108
+ {
109
+ "epoch": 0.03090394025238218,
110
+ "grad_norm": 0.8278765532866457,
111
+ "learning_rate": 3.529411764705883e-05,
112
+ "loss": 0.9063,
113
+ "num_input_tokens_seen": 701808,
114
+ "step": 60
115
+ },
116
+ {
117
+ "epoch": 0.03347926860674736,
118
+ "grad_norm": 0.7285901101792476,
119
+ "learning_rate": 3.8235294117647055e-05,
120
+ "loss": 0.9031,
121
+ "num_input_tokens_seen": 760304,
122
+ "step": 65
123
+ },
124
+ {
125
+ "epoch": 0.036054596961112545,
126
+ "grad_norm": 0.5341783688819233,
127
+ "learning_rate": 4.11764705882353e-05,
128
+ "loss": 0.8991,
129
+ "num_input_tokens_seen": 818760,
130
+ "step": 70
131
+ },
132
+ {
133
+ "epoch": 0.03862992531547772,
134
+ "grad_norm": 0.46059313680988906,
135
+ "learning_rate": 4.411764705882353e-05,
136
+ "loss": 0.9055,
137
+ "num_input_tokens_seen": 877256,
138
+ "step": 75
139
+ },
140
+ {
141
+ "epoch": 0.04120525366984291,
142
+ "grad_norm": 0.8194379237293679,
143
+ "learning_rate": 4.705882352941177e-05,
144
+ "loss": 0.9092,
145
+ "num_input_tokens_seen": 935752,
146
+ "step": 80
147
+ },
148
+ {
149
+ "epoch": 0.043780582024208085,
150
+ "grad_norm": 0.6745093544830881,
151
+ "learning_rate": 5e-05,
152
+ "loss": 0.9069,
153
+ "num_input_tokens_seen": 994216,
154
+ "step": 85
155
+ },
156
+ {
157
+ "epoch": 0.04635591037857327,
158
+ "grad_norm": 0.2894672897884604,
159
+ "learning_rate": 5.294117647058824e-05,
160
+ "loss": 0.8924,
161
+ "num_input_tokens_seen": 1052704,
162
+ "step": 90
163
+ },
164
+ {
165
+ "epoch": 0.04893123873293845,
166
+ "grad_norm": 0.5108489024576455,
167
+ "learning_rate": 5.588235294117647e-05,
168
+ "loss": 0.9059,
169
+ "num_input_tokens_seen": 1111176,
170
+ "step": 95
171
+ },
172
+ {
173
+ "epoch": 0.05150656708730363,
174
+ "grad_norm": 0.40317180386305224,
175
+ "learning_rate": 5.882352941176471e-05,
176
+ "loss": 0.901,
177
+ "num_input_tokens_seen": 1169664,
178
+ "step": 100
179
+ },
180
+ {
181
+ "epoch": 0.05150656708730363,
182
+ "eval_loss": 0.9077914953231812,
183
+ "eval_runtime": 16.8879,
184
+ "eval_samples_per_second": 3.553,
185
+ "eval_steps_per_second": 0.888,
186
+ "num_input_tokens_seen": 1169664,
187
+ "step": 100
188
  }
189
  ],
190
  "logging_steps": 5,
191
  "max_steps": 3400,
192
+ "num_input_tokens_seen": 1169664,
193
  "num_train_epochs": 2,
194
  "save_steps": 50,
195
  "stateful_callbacks": {
 
204
  "attributes": {}
205
  }
206
  },
207
+ "total_flos": 65637304762368.0,
208
  "train_batch_size": 1,
209
  "trial_name": null,
210
  "trial_params": null