ben81828 commited on
Commit
198e688
1 Parent(s): 93fd562

Training in progress, step 2900, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1404e60fabb5103dbe434a7e9d7d0543eaa50c6cb487e2d805559319b22760ea
3
  size 18516456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f49768fb6efa912a563fdf9c78311249a1d28fb66c672197a06fe3282fb5f3cb
3
  size 18516456
last-checkpoint/global_step2899/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f460b267da24112c290816f70f6e7137581679540f12079396e6fa596f227d10
3
+ size 27700976
last-checkpoint/global_step2899/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d69693a67b4ea41309a61a3ffb732cadb9a0af9665e90775ba2140ca9c9ef4d
3
+ size 27700976
last-checkpoint/global_step2899/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd04bb9b2b31222f2a1031a263103260b1efd98e8a6328299f3354c8bd75e208
3
+ size 27700976
last-checkpoint/global_step2899/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23b8c495c7d5697816e1ca24073f934211e838d5ce8d558e296da94a623d57ce
3
+ size 27700976
last-checkpoint/global_step2899/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62f38235c2e23e938a4dd5ea4feae6b61cd97138c092570aeedf2bde303d6784
3
+ size 411571
last-checkpoint/global_step2899/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1e479890dcb45e1d295d4f36f50fd9c05ff83ef783297bca83a8a59ee446f67
3
+ size 411507
last-checkpoint/global_step2899/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4feeeb48e41148c692fda7743fe9c0e24bb7bab99843a4be41bdd24cdbf4b8fc
3
+ size 411507
last-checkpoint/global_step2899/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c68da5f7c173d41b9d207e860c424c6418eb30b60731c2bdf8bb7524c2b86d27
3
+ size 411507
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step2849
 
1
+ global_step2899
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce92cea831a04716b4b472f1dad1cc986b2021dee9aac057217f5d455b27ec42
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfc5d4e344535f1dd0ff5275633ec3d55bb6249e432442365ff24445d82ec35c
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cddb73bbdf0f6f6a2c3182d70f7ad5d587353b164c08dd4f383b940d6b61e4e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a898928042c09dc123c1025557279997043b7f607bc91ee2ff2d4b4d2b9f1ba
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b24b508e466beb446d37377d2a04757d3bc2b4230de3ac56b25a65d7753a74c1
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6923d07d979aa78d66765208f598662fd5092b5227cd87920feedfb923fa375f
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4c6a18a7de8b25b21673ba2ff7efbaaae00ec8c453c7975b467c1df87b87022
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9ba35b9b3c512fbe857d909557329ba47dbefe5f521014123c05901c32edb6d
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28cb8f1d536b1fc196e353ec37fd07d574fab0a464ddd8b31c73c59dcab3c03b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dda654474af0b5740253998af53f5373a7c152f4ee919c2beac18e6e6497944
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.6319106221199036,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1600",
4
- "epoch": 1.467679629152717,
5
  "eval_steps": 50,
6
- "global_step": 2850,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -5080,11 +5080,100 @@
5080
  "eval_steps_per_second": 0.938,
5081
  "num_input_tokens_seen": 33327720,
5082
  "step": 2850
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5083
  }
5084
  ],
5085
  "logging_steps": 5,
5086
  "max_steps": 3400,
5087
- "num_input_tokens_seen": 33327720,
5088
  "num_train_epochs": 2,
5089
  "save_steps": 50,
5090
  "stateful_callbacks": {
@@ -5099,7 +5188,7 @@
5099
  "attributes": {}
5100
  }
5101
  },
5102
- "total_flos": 1871508641415168.0,
5103
  "train_batch_size": 1,
5104
  "trial_name": null,
5105
  "trial_params": null
 
1
  {
2
  "best_metric": 0.6319106221199036,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4_frozenVision/lora/sft/checkpoint-1600",
4
+ "epoch": 1.4934329126963688,
5
  "eval_steps": 50,
6
+ "global_step": 2900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
5080
  "eval_steps_per_second": 0.938,
5081
  "num_input_tokens_seen": 33327720,
5082
  "step": 2850
5083
+ },
5084
+ {
5085
+ "epoch": 1.4702549575070822,
5086
+ "grad_norm": 6.856320486947826,
5087
+ "learning_rate": 6.861743852128233e-06,
5088
+ "loss": 0.2811,
5089
+ "num_input_tokens_seen": 33386160,
5090
+ "step": 2855
5091
+ },
5092
+ {
5093
+ "epoch": 1.4728302858614473,
5094
+ "grad_norm": 8.133776634702407,
5095
+ "learning_rate": 6.7393129075627335e-06,
5096
+ "loss": 0.2394,
5097
+ "num_input_tokens_seen": 33444648,
5098
+ "step": 2860
5099
+ },
5100
+ {
5101
+ "epoch": 1.4754056142158125,
5102
+ "grad_norm": 5.884612144672532,
5103
+ "learning_rate": 6.6179050846274515e-06,
5104
+ "loss": 0.243,
5105
+ "num_input_tokens_seen": 33503144,
5106
+ "step": 2865
5107
+ },
5108
+ {
5109
+ "epoch": 1.4779809425701778,
5110
+ "grad_norm": 7.133095118516192,
5111
+ "learning_rate": 6.497523254635296e-06,
5112
+ "loss": 0.242,
5113
+ "num_input_tokens_seen": 33561600,
5114
+ "step": 2870
5115
+ },
5116
+ {
5117
+ "epoch": 1.4805562709245428,
5118
+ "grad_norm": 3.725193081900286,
5119
+ "learning_rate": 6.37817026463432e-06,
5120
+ "loss": 0.1864,
5121
+ "num_input_tokens_seen": 33620056,
5122
+ "step": 2875
5123
+ },
5124
+ {
5125
+ "epoch": 1.483131599278908,
5126
+ "grad_norm": 5.26408055314188,
5127
+ "learning_rate": 6.25984893734034e-06,
5128
+ "loss": 0.2406,
5129
+ "num_input_tokens_seen": 33678512,
5130
+ "step": 2880
5131
+ },
5132
+ {
5133
+ "epoch": 1.4857069276332733,
5134
+ "grad_norm": 5.139938399894378,
5135
+ "learning_rate": 6.142562071070179e-06,
5136
+ "loss": 0.2287,
5137
+ "num_input_tokens_seen": 33736960,
5138
+ "step": 2885
5139
+ },
5140
+ {
5141
+ "epoch": 1.4882822559876385,
5142
+ "grad_norm": 5.551633292498772,
5143
+ "learning_rate": 6.026312439675552e-06,
5144
+ "loss": 0.2643,
5145
+ "num_input_tokens_seen": 33795416,
5146
+ "step": 2890
5147
+ },
5148
+ {
5149
+ "epoch": 1.4908575843420036,
5150
+ "grad_norm": 5.974549504189433,
5151
+ "learning_rate": 5.911102792477357e-06,
5152
+ "loss": 0.2956,
5153
+ "num_input_tokens_seen": 33853936,
5154
+ "step": 2895
5155
+ },
5156
+ {
5157
+ "epoch": 1.4934329126963688,
5158
+ "grad_norm": 5.786971041370645,
5159
+ "learning_rate": 5.796935854200763e-06,
5160
+ "loss": 0.29,
5161
+ "num_input_tokens_seen": 33912440,
5162
+ "step": 2900
5163
+ },
5164
+ {
5165
+ "epoch": 1.4934329126963688,
5166
+ "eval_loss": 0.8113046884536743,
5167
+ "eval_runtime": 16.0025,
5168
+ "eval_samples_per_second": 3.749,
5169
+ "eval_steps_per_second": 0.937,
5170
+ "num_input_tokens_seen": 33912440,
5171
+ "step": 2900
5172
  }
5173
  ],
5174
  "logging_steps": 5,
5175
  "max_steps": 3400,
5176
+ "num_input_tokens_seen": 33912440,
5177
  "num_train_epochs": 2,
5178
  "save_steps": 50,
5179
  "stateful_callbacks": {
 
5188
  "attributes": {}
5189
  }
5190
  },
5191
+ "total_flos": 1904344826904576.0,
5192
  "train_batch_size": 1,
5193
  "trial_name": null,
5194
  "trial_params": null