imdatta0 commited on
Commit
60ee534
1 Parent(s): 56726a7

Training in progress, step 13

Browse files
adapter_config.json CHANGED
@@ -51,7 +51,7 @@
51
  40
52
  ],
53
  "loftq_config": {},
54
- "lora_alpha": 8,
55
  "lora_dropout": 0,
56
  "megatron_config": null,
57
  "megatron_core": "megatron.core",
@@ -61,13 +61,13 @@
61
  "rank_pattern": {},
62
  "revision": "unsloth",
63
  "target_modules": [
64
- "up_proj",
65
  "gate_proj",
 
66
  "v_proj",
67
- "k_proj",
68
- "o_proj",
69
  "q_proj",
70
- "down_proj"
 
71
  ],
72
  "task_type": "CAUSAL_LM"
73
  }
 
51
  40
52
  ],
53
  "loftq_config": {},
54
+ "lora_alpha": 16,
55
  "lora_dropout": 0,
56
  "megatron_config": null,
57
  "megatron_core": "megatron.core",
 
61
  "rank_pattern": {},
62
  "revision": "unsloth",
63
  "target_modules": [
64
+ "o_proj",
65
  "gate_proj",
66
+ "down_proj",
67
  "v_proj",
 
 
68
  "q_proj",
69
+ "up_proj",
70
+ "k_proj"
71
  ],
72
  "task_type": "CAUSAL_LM"
73
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a374c2b88be072d9716e65bf9e02d9b39b0174f96dbce8c010a7c734696c5123
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faf8e2643c43af54226eaf90c2975f3c64f36a365a5e7a8a533dc69b51e90591
3
  size 83945296
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3540d0fa1d706c9497b7cfe9a65225c70d741238896c851334610d66d61e2d5a
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08aaf988e5d210f12c71423799c079d9355eed05603279a94a92ebcd1afd8cfd
3
  size 5112