noahtye commited on
Commit
126b507
1 Parent(s): c968679

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "down_proj",
25
  "q_proj",
 
26
  "up_proj",
 
27
  "v_proj",
28
- "gate_proj",
29
- "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
23
  "q_proj",
24
+ "k_proj",
25
  "up_proj",
26
+ "down_proj",
27
  "v_proj",
28
+ "o_proj",
29
+ "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48895384cc999a8f83a584c43badd74b8ee4235fce9c96622e66b6055c766e27
3
  size 41581360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a47b01c1852e09c0608608d6bd4ed2a167d68b2866438d3304d5945993823ff
3
  size 41581360
runs/Sep10_23-56-18_2237ab4ecc6a/events.out.tfevents.1726012584.2237ab4ecc6a.1213.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ddc62f60672eae84b50179ae1d8c7516511e3c01be83051c6199440bbac6e88
3
+ size 10104
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f289bc05132635a8bc7aca7aa21255efd5e18f3710f43e3cdb96bcd41be4922
3
- size 17525357
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d12cb475f583be7a862eb1e11c3e558581c2f9a65be8e8a8ea86381ed22f301
3
+ size 17525456
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:abf03143b82ef161153d5deb02ccd959db20f0629e960ecae76aed41f676e372
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a2b4a433844a85382b2ae97075b4c2a9852e8b78dfca963794e72eafda0b566
3
  size 5496