zlsl commited on
Commit
abdd9b2
·
1 Parent(s): 5551b86

Upload 2 files

Browse files
Files changed (2) hide show
  1. adapter_config.json +4 -4
  2. adapter_model.safetensors +2 -2
adapter_config.json CHANGED
@@ -8,15 +8,15 @@
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
  "lora_alpha": 16.0,
11
- "lora_dropout": 0.0,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
- "r": 64,
15
  "revision": null,
16
  "target_modules": [
17
  "c_attn",
18
- "c_fc",
19
- "c_proj"
20
  ],
21
  "task_type": "CAUSAL_LM"
22
  }
 
8
  "layers_pattern": null,
9
  "layers_to_transform": null,
10
  "lora_alpha": 16.0,
11
+ "lora_dropout": 0.05,
12
  "modules_to_save": null,
13
  "peft_type": "LORA",
14
+ "r": 16,
15
  "revision": null,
16
  "target_modules": [
17
  "c_attn",
18
+ "c_proj",
19
+ "c_fc"
20
  ],
21
  "task_type": "CAUSAL_LM"
22
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c39b64160384daf5ef11102c2a97d014b3ce5b70edfe8c42bbf320d59c53c6bc
3
- size 838903032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdaa639a308fc0e6a2da03e59825c364c746bd8d2f417a35271fc9952ffa6e68
3
+ size 209757176