ben81828 commited on
Commit
5656078
·
verified ·
1 Parent(s): eb03109

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e8b8b0d35e964892c4538367f208dcb82426a1165c1fe8d9ec6ab99f0299952
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87a05b954f6ad396e106053a5bf73274eb4671d1ca4b7518421076c40296fc81
3
  size 29034840
last-checkpoint/global_step750/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fdf3a0693bea36f402ce1ba4fe57537b0823ae9ed524dc7d6f915dc67c85dd8
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eaba375ba8e10fb1d10e83cbcdc6fd00f9fee9f58f0458c57c0bd8732cb80a5d
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec3c4abbc30f92b4860a5c61d44fd2577b8365be30936d562e293a745076e7cd
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9753134f6593118672c497761e949545267f45e52527068e5483550bd2453f6a
3
+ size 43429616
last-checkpoint/global_step750/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7ceefcf5f9aaf423f675a59b70c7f49dd63bf2969dd0f241c706afca330069
3
+ size 637299
last-checkpoint/global_step750/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88f0e9d3d0064652385de4668d7d1493464ea47cc90272985cc5f82378fe9407
3
+ size 637171
last-checkpoint/global_step750/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2747d5b9a430ea06aa5e313062964f76ced72bafedfb86736e283a3440555b
3
+ size 637171
last-checkpoint/global_step750/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df739d3279f7e7c62c77653c661bc9074781cffe48c2820c6dcb897ac09d46b
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step700
 
1
+ global_step750
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7d74de51245105e1fbf57a6707ef3538b353952485508f6e2f8f74dc5d479d4
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66b4ef73f3603a1b91082ee108fa8299ebe45fb3cdeec7d0bdca1982af5bf07d
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0617c9eb6cf7df57b2e0bb53cfe17c05f0910de56fe5b14427fe39ab54a44782
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:882eccb2a272cf97cd645050bd077c971e48e78584f717a1b1cc9b5f1c9326dc
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed68a365057022897d9645ee60902a77102f43215dcdf2ddd5d3842b6a8446d8
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1626bac54f5806a604b7efdd749c5b65d63bbb40fc55c3744aae6130aa24f3de
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63ebaa0c302cadbdfcd9f8ee2289e35ecf9c9fc8c9968fc0c05f100dac20c6b9
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40cf83ff997228172cf0b991f9d5209728ccf2f0a75841db5e31e647779a1ad2
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:01460e95e3e56a90233dbd46a8dfe9605b67858d03a9beec52437f5ab98bbf42
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb90ddaf9f2555f7a751dcba5b295eed1e25f610d0357becbe825d3cf6fda52d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.5883122682571411,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_direction_then_DetectAndClassify_scale6/lora/sft/checkpoint-600",
4
- "epoch": 0.206794682422452,
5
  "eval_steps": 50,
6
- "global_step": 700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1253,11 +1253,100 @@
1253
  "eval_steps_per_second": 0.785,
1254
  "num_input_tokens_seen": 7262832,
1255
  "step": 700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1256
  }
1257
  ],
1258
  "logging_steps": 5,
1259
  "max_steps": 6770,
1260
- "num_input_tokens_seen": 7262832,
1261
  "num_train_epochs": 2,
1262
  "save_steps": 50,
1263
  "stateful_callbacks": {
@@ -1272,7 +1361,7 @@
1272
  "attributes": {}
1273
  }
1274
  },
1275
- "total_flos": 479089626185728.0,
1276
  "train_batch_size": 1,
1277
  "trial_name": null,
1278
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.4665524661540985,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_direction_then_DetectAndClassify_scale6/lora/sft/checkpoint-750",
4
+ "epoch": 0.22156573116691286,
5
  "eval_steps": 50,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1253
  "eval_steps_per_second": 0.785,
1254
  "num_input_tokens_seen": 7262832,
1255
  "step": 700
1256
+ },
1257
+ {
1258
+ "epoch": 0.20827178729689808,
1259
+ "grad_norm": 5.1484150923671,
1260
+ "learning_rate": 9.92029461479909e-05,
1261
+ "loss": 0.5151,
1262
+ "num_input_tokens_seen": 7314520,
1263
+ "step": 705
1264
+ },
1265
+ {
1266
+ "epoch": 0.20974889217134415,
1267
+ "grad_norm": 10.902688361325707,
1268
+ "learning_rate": 9.918108001438283e-05,
1269
+ "loss": 0.6158,
1270
+ "num_input_tokens_seen": 7365368,
1271
+ "step": 710
1272
+ },
1273
+ {
1274
+ "epoch": 0.21122599704579026,
1275
+ "grad_norm": 6.960249994011121,
1276
+ "learning_rate": 9.915892046694464e-05,
1277
+ "loss": 0.5164,
1278
+ "num_input_tokens_seen": 7417296,
1279
+ "step": 715
1280
+ },
1281
+ {
1282
+ "epoch": 0.21270310192023634,
1283
+ "grad_norm": 6.138105593354917,
1284
+ "learning_rate": 9.913646763787992e-05,
1285
+ "loss": 0.5823,
1286
+ "num_input_tokens_seen": 7469640,
1287
+ "step": 720
1288
+ },
1289
+ {
1290
+ "epoch": 0.21418020679468242,
1291
+ "grad_norm": 5.865897561310743,
1292
+ "learning_rate": 9.911372166114208e-05,
1293
+ "loss": 0.5145,
1294
+ "num_input_tokens_seen": 7521520,
1295
+ "step": 725
1296
+ },
1297
+ {
1298
+ "epoch": 0.21565731166912852,
1299
+ "grad_norm": 14.330290158050401,
1300
+ "learning_rate": 9.909068267243336e-05,
1301
+ "loss": 0.571,
1302
+ "num_input_tokens_seen": 7573880,
1303
+ "step": 730
1304
+ },
1305
+ {
1306
+ "epoch": 0.2171344165435746,
1307
+ "grad_norm": 6.90669118248274,
1308
+ "learning_rate": 9.906735080920413e-05,
1309
+ "loss": 0.4638,
1310
+ "num_input_tokens_seen": 7625896,
1311
+ "step": 735
1312
+ },
1313
+ {
1314
+ "epoch": 0.21861152141802068,
1315
+ "grad_norm": 6.310395883326308,
1316
+ "learning_rate": 9.904372621065206e-05,
1317
+ "loss": 0.5449,
1318
+ "num_input_tokens_seen": 7676528,
1319
+ "step": 740
1320
+ },
1321
+ {
1322
+ "epoch": 0.22008862629246675,
1323
+ "grad_norm": 27.446565035007364,
1324
+ "learning_rate": 9.901980901772126e-05,
1325
+ "loss": 0.5505,
1326
+ "num_input_tokens_seen": 7728240,
1327
+ "step": 745
1328
+ },
1329
+ {
1330
+ "epoch": 0.22156573116691286,
1331
+ "grad_norm": 5.6934624405279655,
1332
+ "learning_rate": 9.899559937310148e-05,
1333
+ "loss": 0.514,
1334
+ "num_input_tokens_seen": 7779872,
1335
+ "step": 750
1336
+ },
1337
+ {
1338
+ "epoch": 0.22156573116691286,
1339
+ "eval_loss": 0.4665524661540985,
1340
+ "eval_runtime": 19.1629,
1341
+ "eval_samples_per_second": 3.131,
1342
+ "eval_steps_per_second": 0.783,
1343
+ "num_input_tokens_seen": 7779872,
1344
+ "step": 750
1345
  }
1346
  ],
1347
  "logging_steps": 5,
1348
  "max_steps": 6770,
1349
+ "num_input_tokens_seen": 7779872,
1350
  "num_train_epochs": 2,
1351
  "save_steps": 50,
1352
  "stateful_callbacks": {
 
1361
  "attributes": {}
1362
  }
1363
  },
1364
+ "total_flos": 513213288415232.0,
1365
  "train_batch_size": 1,
1366
  "trial_name": null,
1367
  "trial_params": null