liuylhf commited on
Commit
f7592cd
1 Parent(s): 62e483d

Training in progress, step 15

Browse files
adapter_config.json CHANGED
@@ -21,12 +21,12 @@
21
  "revision": null,
22
  "target_modules": [
23
  "k_proj",
24
- "q_proj",
25
  "up_proj",
 
 
26
  "v_proj",
27
  "o_proj",
28
- "down_proj",
29
- "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
21
  "revision": null,
22
  "target_modules": [
23
  "k_proj",
 
24
  "up_proj",
25
+ "gate_proj",
26
+ "down_proj",
27
  "v_proj",
28
  "o_proj",
29
+ "q_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:016e62df04ec65529a38dd588f716fdaecf1b7d063f449f84b3291580dff92f2
3
  size 828527688
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecff1d71776426f7277227c80b2620c6cc122f1f38cd7fef2e8fbe1944c8dc22
3
  size 828527688
config.json CHANGED
@@ -37,7 +37,7 @@
37
  "rope_theta": 500000.0,
38
  "tie_word_embeddings": false,
39
  "torch_dtype": "bfloat16",
40
- "transformers_version": "4.40.0.dev0",
41
  "use_cache": false,
42
  "vocab_size": 128256
43
  }
 
37
  "rope_theta": 500000.0,
38
  "tie_word_embeddings": false,
39
  "torch_dtype": "bfloat16",
40
+ "transformers_version": "4.40.0",
41
  "use_cache": false,
42
  "vocab_size": 128256
43
  }
runs/Apr25_04-30-32_training-01/events.out.tfevents.1714019438.training-01.52176.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75069b385b5d3f6326a490eb622f67fe68ce02fe72b207d58c8e1ba47c588fdb
3
+ size 10012
tokenizer.json CHANGED
@@ -2329,10 +2329,69 @@
2329
  ]
2330
  },
2331
  "post_processor": {
2332
- "type": "ByteLevel",
2333
- "add_prefix_space": true,
2334
- "trim_offsets": false,
2335
- "use_regex": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2336
  },
2337
  "decoder": {
2338
  "type": "ByteLevel",
@@ -2348,6 +2407,7 @@
2348
  "end_of_word_suffix": null,
2349
  "fuse_unk": false,
2350
  "byte_fallback": false,
 
2351
  "vocab": {
2352
  "!": 0,
2353
  "\"": 1,
 
2329
  ]
2330
  },
2331
  "post_processor": {
2332
+ "type": "Sequence",
2333
+ "processors": [
2334
+ {
2335
+ "type": "ByteLevel",
2336
+ "add_prefix_space": true,
2337
+ "trim_offsets": false,
2338
+ "use_regex": true
2339
+ },
2340
+ {
2341
+ "type": "TemplateProcessing",
2342
+ "single": [
2343
+ {
2344
+ "SpecialToken": {
2345
+ "id": "<|begin_of_text|>",
2346
+ "type_id": 0
2347
+ }
2348
+ },
2349
+ {
2350
+ "Sequence": {
2351
+ "id": "A",
2352
+ "type_id": 0
2353
+ }
2354
+ }
2355
+ ],
2356
+ "pair": [
2357
+ {
2358
+ "SpecialToken": {
2359
+ "id": "<|begin_of_text|>",
2360
+ "type_id": 0
2361
+ }
2362
+ },
2363
+ {
2364
+ "Sequence": {
2365
+ "id": "A",
2366
+ "type_id": 0
2367
+ }
2368
+ },
2369
+ {
2370
+ "SpecialToken": {
2371
+ "id": "<|begin_of_text|>",
2372
+ "type_id": 1
2373
+ }
2374
+ },
2375
+ {
2376
+ "Sequence": {
2377
+ "id": "B",
2378
+ "type_id": 1
2379
+ }
2380
+ }
2381
+ ],
2382
+ "special_tokens": {
2383
+ "<|begin_of_text|>": {
2384
+ "id": "<|begin_of_text|>",
2385
+ "ids": [
2386
+ 128000
2387
+ ],
2388
+ "tokens": [
2389
+ "<|begin_of_text|>"
2390
+ ]
2391
+ }
2392
+ }
2393
+ }
2394
+ ]
2395
  },
2396
  "decoder": {
2397
  "type": "ByteLevel",
 
2407
  "end_of_word_suffix": null,
2408
  "fuse_unk": false,
2409
  "byte_fallback": false,
2410
+ "ignore_merges": true,
2411
  "vocab": {
2412
  "!": 0,
2413
  "\"": 1,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44b5f3a9294b75563476b806eee347de0294a1f6b9784bc9e9cf15233efba500
3
  size 6328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33c270181dae82aff969aff9c90f69d7dc1388ef1fffb395c4a9a45efca78d06
3
  size 6328