ben81828 commited on
Commit
2b7564d
·
verified ·
1 Parent(s): fcbd1b2

Training in progress, step 1200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:050dc1f1854d68a42fb308d9dd143ddc3d6fc0d86870ea1eac6bf90317df973a
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419197c7b895b88c7afacbe056bbf0e1d22ab3d4ba5caafffba674acfa991533
3
  size 29034840
last-checkpoint/global_step1200/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb1aec5f88c70086eb769372a167d1660968748cc110327de5c320debe43f371
3
+ size 43429616
last-checkpoint/global_step1200/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae158e05ebda4f41c8ce55623f0bddf3afa6bbc045c2ada936c7da24035799d3
3
+ size 43429616
last-checkpoint/global_step1200/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26f6bf14fc9c2c40db373922da15d23c0d08a53dc9d0c74f5f03dc1437ca0fb7
3
+ size 43429616
last-checkpoint/global_step1200/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31febf6ba053a4152cb24e518f2e806dda3c3010aa10d75d7e5a858d33d3eb36
3
+ size 43429616
last-checkpoint/global_step1200/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d06f5911185a6ebb03a50cfa88a0b397d7dcb732800b879d25a0b602da1df67
3
+ size 637299
last-checkpoint/global_step1200/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ddb6eccf101a9bfdd7a0e209ad8e34c5bd436c139a6cc24e86858cfb663649f
3
+ size 637171
last-checkpoint/global_step1200/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63fa0ba1e22cd4330a732b53b080494fee16b561a8509fb226cb0abdd186e7ea
3
+ size 637171
last-checkpoint/global_step1200/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a54361c50a0229f80835738434c247594a5157c91a95390f6abdef0e295dfb9d
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1150
 
1
+ global_step1200
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:274dc3860ee0c7f4d5348f60910a4b568498c04adfefb89f905b1c78a82c1312
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a209a0c0025f9ce8e2beeba50c1f0828d5c34a2482310fcd0bf5fc24c2c67be2
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9aa441491b9ca89e796944520fa1db332a67c0a1a920be83edd2d96d741716d
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a67fb929b8c51f9b1c6ff9f11366e57e55128a1d36df85a9d37a008b49017a75
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1ee3434533b24fb771504fa8cceb5c2ea25fe0de1641128feaceccc65afe6ed
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b9ef3b0c0978d0b611f4257c939f1c2c6f07e6227bfea6675532d285b0b64a7
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93b4a44be1335173d2e3120bd0d1e6346f3e832d8935752c70ce1e98f017fa87
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0183d14c8ed52ee533139532e9bcf7bc34ec297a064845b35741cb501d92675f
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d461c8d7517d4b88333bff7984fc3bfc149292198b04bbc18a49aee698ffb5c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16e0cffc6b063574ed312ee2198c86d3dddf2450d0400f042fdc08bd22dcbc7f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.5566551155643538e-05,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_scale4/lora/sft/checkpoint-1150",
4
- "epoch": 0.5923255215039918,
5
  "eval_steps": 50,
6
- "global_step": 1150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -2054,11 +2054,100 @@
2054
  "eval_steps_per_second": 0.775,
2055
  "num_input_tokens_seen": 11481600,
2056
  "step": 1150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2057
  }
2058
  ],
2059
  "logging_steps": 5,
2060
  "max_steps": 3400,
2061
- "num_input_tokens_seen": 11481600,
2062
  "num_train_epochs": 2,
2063
  "save_steps": 50,
2064
  "stateful_callbacks": {
@@ -2073,7 +2162,7 @@
2073
  "attributes": {}
2074
  }
2075
  },
2076
- "total_flos": 757564358524928.0,
2077
  "train_batch_size": 1,
2078
  "trial_name": null,
2079
  "trial_params": null
 
1
  {
2
  "best_metric": 1.5566551155643538e-05,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_direction_scale4/lora/sft/checkpoint-1150",
4
+ "epoch": 0.6180788050476436,
5
  "eval_steps": 50,
6
+ "global_step": 1200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
2054
  "eval_steps_per_second": 0.775,
2055
  "num_input_tokens_seen": 11481600,
2056
  "step": 1150
2057
+ },
2058
+ {
2059
+ "epoch": 0.5949008498583569,
2060
+ "grad_norm": 0.00046149068264744525,
2061
+ "learning_rate": 7.875623046909544e-05,
2062
+ "loss": 0.0003,
2063
+ "num_input_tokens_seen": 11531520,
2064
+ "step": 1155
2065
+ },
2066
+ {
2067
+ "epoch": 0.5974761782127221,
2068
+ "grad_norm": 0.0007891686440962277,
2069
+ "learning_rate": 7.855697251155967e-05,
2070
+ "loss": 0.0,
2071
+ "num_input_tokens_seen": 11581440,
2072
+ "step": 1160
2073
+ },
2074
+ {
2075
+ "epoch": 0.6000515065670873,
2076
+ "grad_norm": 6.055667685356653,
2077
+ "learning_rate": 7.835703917741212e-05,
2078
+ "loss": 0.0414,
2079
+ "num_input_tokens_seen": 11631360,
2080
+ "step": 1165
2081
+ },
2082
+ {
2083
+ "epoch": 0.6026268349214525,
2084
+ "grad_norm": 3.9367530795344696,
2085
+ "learning_rate": 7.81564351951057e-05,
2086
+ "loss": 0.02,
2087
+ "num_input_tokens_seen": 11681280,
2088
+ "step": 1170
2089
+ },
2090
+ {
2091
+ "epoch": 0.6052021632758177,
2092
+ "grad_norm": 0.014148361508830113,
2093
+ "learning_rate": 7.795516530895414e-05,
2094
+ "loss": 0.0002,
2095
+ "num_input_tokens_seen": 11731200,
2096
+ "step": 1175
2097
+ },
2098
+ {
2099
+ "epoch": 0.6077774916301828,
2100
+ "grad_norm": 0.12219412030414617,
2101
+ "learning_rate": 7.775323427901993e-05,
2102
+ "loss": 0.0026,
2103
+ "num_input_tokens_seen": 11781120,
2104
+ "step": 1180
2105
+ },
2106
+ {
2107
+ "epoch": 0.610352819984548,
2108
+ "grad_norm": 5.108798582105785,
2109
+ "learning_rate": 7.755064688100171e-05,
2110
+ "loss": 0.0168,
2111
+ "num_input_tokens_seen": 11831040,
2112
+ "step": 1185
2113
+ },
2114
+ {
2115
+ "epoch": 0.6129281483389132,
2116
+ "grad_norm": 0.004932933139070266,
2117
+ "learning_rate": 7.734740790612136e-05,
2118
+ "loss": 0.0037,
2119
+ "num_input_tokens_seen": 11880960,
2120
+ "step": 1190
2121
+ },
2122
+ {
2123
+ "epoch": 0.6155034766932784,
2124
+ "grad_norm": 0.022109282268939556,
2125
+ "learning_rate": 7.714352216101055e-05,
2126
+ "loss": 0.0007,
2127
+ "num_input_tokens_seen": 11930880,
2128
+ "step": 1195
2129
+ },
2130
+ {
2131
+ "epoch": 0.6180788050476436,
2132
+ "grad_norm": 0.037426371340479145,
2133
+ "learning_rate": 7.693899446759727e-05,
2134
+ "loss": 0.0038,
2135
+ "num_input_tokens_seen": 11980800,
2136
+ "step": 1200
2137
+ },
2138
+ {
2139
+ "epoch": 0.6180788050476436,
2140
+ "eval_loss": 0.0022160401567816734,
2141
+ "eval_runtime": 18.725,
2142
+ "eval_samples_per_second": 3.204,
2143
+ "eval_steps_per_second": 0.801,
2144
+ "num_input_tokens_seen": 11980800,
2145
+ "step": 1200
2146
  }
2147
  ],
2148
  "logging_steps": 5,
2149
  "max_steps": 3400,
2150
+ "num_input_tokens_seen": 11980800,
2151
  "num_train_epochs": 2,
2152
  "save_steps": 50,
2153
  "stateful_callbacks": {
 
2162
  "attributes": {}
2163
  }
2164
  },
2165
+ "total_flos": 790504338096128.0,
2166
  "train_batch_size": 1,
2167
  "trial_name": null,
2168
  "trial_params": null