zeren commited on
Commit
bf6a65c
1 Parent(s): 09422d5

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -17,18 +17,6 @@
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
  "pretraining_tp": 1,
20
- "quantization_config": {
21
- "bnb_4bit_compute_dtype": "bfloat16",
22
- "bnb_4bit_quant_type": "nf4",
23
- "bnb_4bit_use_double_quant": true,
24
- "llm_int8_enable_fp32_cpu_offload": false,
25
- "llm_int8_has_fp16_weight": false,
26
- "llm_int8_skip_modules": null,
27
- "llm_int8_threshold": 6.0,
28
- "load_in_4bit": true,
29
- "load_in_8bit": false,
30
- "quant_method": "bitsandbytes"
31
- },
32
  "rms_norm_eps": 1e-05,
33
  "rope_scaling": null,
34
  "rope_theta": 500000.0,
 
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
20
  "rms_norm_eps": 1e-05,
21
  "rope_scaling": null,
22
  "rope_theta": 500000.0,
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c32fe64099bf085e8ed7a61f905b088f611e971e50982cee15bedbee54cfef2
3
+ size 4976698672
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61bdbad45969674db1b07e660e9c1da815e6b7f0ae51f04b9284f32198c2a1d0
3
+ size 4999802720
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9488998632e6a1cac71dbec02a8bcd5dd8ee3c36cd7a827dfd1d025a811d2468
3
+ size 4915916176
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2873dd0818cedf424b2db711048c77c259e45b021c5ff0769a1acebedea798b
3
+ size 1168138808
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff