bhuvanmdev commited on
Commit
2bfe9b6
1 Parent(s): 80b16c6

Training in progress, step 1400, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -20,9 +20,9 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "qkv_proj",
24
  "down_proj",
25
- "o_proj",
26
  "gate_up_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "o_proj",
24
  "qkv_proj",
25
  "down_proj",
 
26
  "gate_up_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6b616fe2eb021d20d4875636738be3ca7342fc6db866fd391eb53743dd565cb0
3
  size 100697728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:070555ab40db51d3a5840454fa269b784df6d102a3f8624f43e74d58478c8951
3
  size 100697728
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5356f6af8453685dfcd2ab29084ffd177cb62b8eaee4a3289a76f99910495ddc
3
  size 201541754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ca946250efe75fef35aba1dbf1b03c113298722b5d7da5cedd58c3eb48dddaf
3
  size 201541754
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44f04742b905a4ba4f17b43b2fddc49db3e772b856badc745872559eec73cca5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:354c3e84f6b2a094e741c8f1f43bff92126887859b832aecb38252864730d160
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc40c38146bda7179f03752fd8672ec7d4feef1519e7b61e194eda1f35f35156
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded0d7f7cfcac928a75f77d93dbcaf43644e4251cba1e72644318143074c5e26
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4937388193202147,
5
  "eval_steps": 500,
6
- "global_step": 1380,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1111,14 +1111,30 @@
1111
  "loss": 0.4092,
1112
  "num_input_tokens_seen": 928645,
1113
  "step": 1380
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1114
  }
1115
  ],
1116
  "logging_steps": 10,
1117
  "max_steps": 2795,
1118
- "num_input_tokens_seen": 928645,
1119
  "num_train_epochs": 1,
1120
  "save_steps": 20,
1121
- "total_flos": 2.088194664047616e+16,
1122
  "train_batch_size": 1,
1123
  "trial_name": null,
1124
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.5008944543828264,
5
  "eval_steps": 500,
6
+ "global_step": 1400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1111
  "loss": 0.4092,
1112
  "num_input_tokens_seen": 928645,
1113
  "step": 1380
1114
+ },
1115
+ {
1116
+ "epoch": 0.49731663685152055,
1117
+ "grad_norm": 0.3198321759700775,
1118
+ "learning_rate": 0.00010053667262969588,
1119
+ "loss": 0.3814,
1120
+ "num_input_tokens_seen": 933801,
1121
+ "step": 1390
1122
+ },
1123
+ {
1124
+ "epoch": 0.5008944543828264,
1125
+ "grad_norm": 0.3592207133769989,
1126
+ "learning_rate": 9.982110912343471e-05,
1127
+ "loss": 0.3776,
1128
+ "num_input_tokens_seen": 938905,
1129
+ "step": 1400
1130
  }
1131
  ],
1132
  "logging_steps": 10,
1133
  "max_steps": 2795,
1134
+ "num_input_tokens_seen": 938905,
1135
  "num_train_epochs": 1,
1136
  "save_steps": 20,
1137
+ "total_flos": 2.111265780839424e+16,
1138
  "train_batch_size": 1,
1139
  "trial_name": null,
1140
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8bee6a61298e4ba0636e6bf836a1f169e31db19f42b2bb480b9f85ccd38a9c1e
3
  size 5048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77d0a7fc1510ab3970debab54d5361ccc4e46d17c58f6086837651cc376b2329
3
  size 5048