ben81828 commited on
Commit
055534d
·
verified ·
1 Parent(s): 7e264f0

Training in progress, step 650, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1513791e52117a6d639f05fdddf98b812e9b7c86a4830e79581dc7ae45e41a8
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ae4ae60a26ee99e11a2e5ab6fb595036c6ed6a11a5c18815e41835fd861d1bc
3
  size 18516456
last-checkpoint/global_step650/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9eef2952c199856a9fcf62051a5b2ab7c2f8370321900cbcaae7ed004133b4c4
3
+ size 27700976
last-checkpoint/global_step650/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db74a445318eaaff0a6999434cc220fb4f7ccc28d808a5796953bcfcee002d5a
3
+ size 27700976
last-checkpoint/global_step650/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ece5498497c6a7ba1102982002c1c274bb424b8a8bacb73b0653ec4d2d4bca6a
3
+ size 27700976
last-checkpoint/global_step650/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24f8192475e214e8f6683656a8f114c5e3c53003c057186a9b5628689b634ba7
3
+ size 27700976
last-checkpoint/global_step650/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7d2a717f34038e4b5233e0e7ad841d4c46a075d159b8407dd5f2e81acde48dd
3
+ size 411571
last-checkpoint/global_step650/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5993695ce009775d7279eea0388f12ae416392537b3d54d2b2f8f3b716ffabc9
3
+ size 411507
last-checkpoint/global_step650/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d77f3fb1b5c101cecf264192e8b86fc1c3f2ecf4f47550830eab354ff674beb3
3
+ size 411507
last-checkpoint/global_step650/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeae3481e62915481d3b6289b00da7b260e3fa552df9978394109f43cd099b42
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step600
 
1
+ global_step650
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a81e3916b1392c4c49afb171dee5415c15f5a5a5af8749b28195fcfa0596699c
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8044e4c53158c210a17648ba8f2dc2d25a25bbfc55f686015542618eb652a33e
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a781038dd714b87b8adb1aac8dbc8217ceb607428a992133954ad522365236e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cd85d7fa425e7888c973f1c2985ac15ca21b5e6171fe140a401c2bc75ca46ff
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9446c3db15f382a5546f13622787fc99392a5e0bc8a9ca2da1838de7ab621a37
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7915667371a58f1598639e0d1c20a0c59c783c14580cd040a6631eb4ea2311e
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f11e7a6b3faa884fc23044e3772ff9dd72c257f02e121665061e2a03d518bd9
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35dd78929ad7f0fbf37fdb1284e8edf0424350f6e6ce1cd5a3ee78979af3d3cb
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b76b388bede074656df32b92902ac42b965557bfee0c930366af07d8382b1b4a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd94ecf5c982ee0e060d3e07a575ce03dc3b0f289b5e32a1f65d3b6366a8a0e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.8908902406692505,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-300",
4
- "epoch": 0.3090394025238218,
5
  "eval_steps": 50,
6
- "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1075,11 +1075,100 @@
1075
  "eval_steps_per_second": 0.876,
1076
  "num_input_tokens_seen": 7017576,
1077
  "step": 600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1078
  }
1079
  ],
1080
  "logging_steps": 5,
1081
  "max_steps": 3400,
1082
- "num_input_tokens_seen": 7017576,
1083
  "num_train_epochs": 2,
1084
  "save_steps": 50,
1085
  "stateful_callbacks": {
@@ -1094,7 +1183,7 @@
1094
  "attributes": {}
1095
  }
1096
  },
1097
- "total_flos": 394023670644736.0,
1098
  "train_batch_size": 1,
1099
  "trial_name": null,
1100
  "trial_params": null
 
1
  {
2
  "best_metric": 0.8908902406692505,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-300",
4
+ "epoch": 0.3347926860674736,
5
  "eval_steps": 50,
6
+ "global_step": 650,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1075
  "eval_steps_per_second": 0.876,
1076
  "num_input_tokens_seen": 7017576,
1077
  "step": 600
1078
+ },
1079
+ {
1080
+ "epoch": 0.311614730878187,
1081
+ "grad_norm": 0.3545950949033049,
1082
+ "learning_rate": 9.55911532374151e-05,
1083
+ "loss": 0.9019,
1084
+ "num_input_tokens_seen": 7076032,
1085
+ "step": 605
1086
+ },
1087
+ {
1088
+ "epoch": 0.31419005923255217,
1089
+ "grad_norm": 0.2355627006333952,
1090
+ "learning_rate": 9.549077815930636e-05,
1091
+ "loss": 0.8956,
1092
+ "num_input_tokens_seen": 7134536,
1093
+ "step": 610
1094
+ },
1095
+ {
1096
+ "epoch": 0.31676538758691736,
1097
+ "grad_norm": 0.17552483625655946,
1098
+ "learning_rate": 9.538932721758474e-05,
1099
+ "loss": 0.898,
1100
+ "num_input_tokens_seen": 7193032,
1101
+ "step": 615
1102
+ },
1103
+ {
1104
+ "epoch": 0.3193407159412825,
1105
+ "grad_norm": 0.1749010635522076,
1106
+ "learning_rate": 9.528680281157999e-05,
1107
+ "loss": 0.8991,
1108
+ "num_input_tokens_seen": 7251568,
1109
+ "step": 620
1110
+ },
1111
+ {
1112
+ "epoch": 0.3219160442956477,
1113
+ "grad_norm": 0.19885182954224315,
1114
+ "learning_rate": 9.518320736600943e-05,
1115
+ "loss": 0.8961,
1116
+ "num_input_tokens_seen": 7310072,
1117
+ "step": 625
1118
+ },
1119
+ {
1120
+ "epoch": 0.3244913726500129,
1121
+ "grad_norm": 0.4778756508206831,
1122
+ "learning_rate": 9.507854333092063e-05,
1123
+ "loss": 0.8994,
1124
+ "num_input_tokens_seen": 7368560,
1125
+ "step": 630
1126
+ },
1127
+ {
1128
+ "epoch": 0.32706670100437807,
1129
+ "grad_norm": 0.4123272743887767,
1130
+ "learning_rate": 9.497281318163346e-05,
1131
+ "loss": 0.8925,
1132
+ "num_input_tokens_seen": 7427040,
1133
+ "step": 635
1134
+ },
1135
+ {
1136
+ "epoch": 0.32964202935874326,
1137
+ "grad_norm": 0.34409942667705734,
1138
+ "learning_rate": 9.486601941868154e-05,
1139
+ "loss": 0.9087,
1140
+ "num_input_tokens_seen": 7485552,
1141
+ "step": 640
1142
+ },
1143
+ {
1144
+ "epoch": 0.3322173577131084,
1145
+ "grad_norm": 0.43327107411223276,
1146
+ "learning_rate": 9.475816456775313e-05,
1147
+ "loss": 0.8924,
1148
+ "num_input_tokens_seen": 7544040,
1149
+ "step": 645
1150
+ },
1151
+ {
1152
+ "epoch": 0.3347926860674736,
1153
+ "grad_norm": 0.6643023904352003,
1154
+ "learning_rate": 9.464925117963133e-05,
1155
+ "loss": 0.904,
1156
+ "num_input_tokens_seen": 7602512,
1157
+ "step": 650
1158
+ },
1159
+ {
1160
+ "epoch": 0.3347926860674736,
1161
+ "eval_loss": 0.90328449010849,
1162
+ "eval_runtime": 16.1444,
1163
+ "eval_samples_per_second": 3.716,
1164
+ "eval_steps_per_second": 0.929,
1165
+ "num_input_tokens_seen": 7602512,
1166
+ "step": 650
1167
  }
1168
  ],
1169
  "logging_steps": 5,
1170
  "max_steps": 3400,
1171
+ "num_input_tokens_seen": 7602512,
1172
  "num_train_epochs": 2,
1173
  "save_steps": 50,
1174
  "stateful_callbacks": {
 
1183
  "attributes": {}
1184
  }
1185
  },
1186
+ "total_flos": 426870884270080.0,
1187
  "train_batch_size": 1,
1188
  "trial_name": null,
1189
  "trial_params": null