NicoNico commited on
Commit
a79ebd9
1 Parent(s): cf4c823

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -4,9 +4,9 @@ tags:
4
  - mlx
5
  ---
6
 
7
- # GreenBitAI/01-Yi-6B-chat-layer-mix-bpw-2.2-mlx
8
- This quantized low-bit model was converted to MLX format from [`GreenBitAI/01-Yi-6B-chat-layer-mix-bpw-2.2`]().
9
- Refer to the [original model card](https://huggingface.co/GreenBitAI/01-Yi-6B-chat-layer-mix-bpw-2.2) for more details on the model.
10
  ## Use with mlx
11
 
12
  ```bash
@@ -16,6 +16,6 @@ pip install gbx-lm
16
  ```python
17
  from gbx_lm import load, generate
18
 
19
- model, tokenizer = load("GreenBitAI/01-Yi-6B-chat-layer-mix-bpw-2.2-mlx")
20
  response = generate(model, tokenizer, prompt="hello", verbose=True)
21
  ```
 
4
  - mlx
5
  ---
6
 
7
+ # GreenBitAI/01-Yi-6B-Chat-layer-mix-bpw-2.2-mlx
8
+ This quantized low-bit model was converted to MLX format from [`GreenBitAI/01-Yi-6B-Chat-layer-mix-bpw-2.2`]().
9
+ Refer to the [original model card](https://huggingface.co/GreenBitAI/01-Yi-6B-Chat-layer-mix-bpw-2.2) for more details on the model.
10
  ## Use with mlx
11
 
12
  ```bash
 
16
  ```python
17
  from gbx_lm import load, generate
18
 
19
+ model, tokenizer = load("GreenBitAI/01-Yi-6B-Chat-layer-mix-bpw-2.2-mlx")
20
  response = generate(model, tokenizer, prompt="hello", verbose=True)
21
  ```
config.json CHANGED
@@ -77,7 +77,7 @@
77
  "top_p": 1.0,
78
  "torch_dtype": "float16",
79
  "torchscript": false,
80
- "transformers_version": "4.39.0.dev0",
81
  "typical_p": 1.0,
82
  "use_bfloat16": false,
83
  "use_cache": true,
 
77
  "top_p": 1.0,
78
  "torch_dtype": "float16",
79
  "torchscript": false,
80
+ "transformers_version": "4.39.3",
81
  "typical_p": 1.0,
82
  "use_bfloat16": false,
83
  "use_cache": true,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8719d9812ae0a98010747e74769fc95f53fa5b2c17b3074219f99a1ce310dc20
3
- size 2868359625
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c7c8e4b071f59e95c8ca9fbc739394c897f070788dba38473bcc06f52591a6d
3
+ size 2867900553
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 2868224000
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 2867765248
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model.safetensors",
quant_strategy.json CHANGED
@@ -89,8 +89,8 @@
89
  }
90
  },
91
  "model.layers.1": {
92
- "accuracy": 0.9216856956481934,
93
- "total_bits": 486917664,
94
  "q_proj": {
95
  "group_size": {
96
  "2": 64
@@ -129,10 +129,10 @@
129
  },
130
  "o_proj": {
131
  "group_size": {
132
- "4": 128
133
  },
134
  "bits": [
135
- 4
136
  ],
137
  "bits_prop": [
138
  1
@@ -165,10 +165,10 @@
165
  },
166
  "down_proj": {
167
  "group_size": {
168
- "4": 128
169
  },
170
  "bits": [
171
- 4
172
  ],
173
  "bits_prop": [
174
  1
@@ -265,8 +265,8 @@
265
  }
266
  },
267
  "model.layers.3": {
268
- "accuracy": 0.9084997177124023,
269
- "total_bits": 449761824,
270
  "q_proj": {
271
  "group_size": {
272
  "2": 64
@@ -341,10 +341,10 @@
341
  },
342
  "down_proj": {
343
  "group_size": {
344
- "4": 128
345
  },
346
  "bits": [
347
- 4
348
  ],
349
  "bits_prop": [
350
  1
@@ -705,8 +705,8 @@
705
  }
706
  },
707
  "model.layers.8": {
708
- "accuracy": 0.9228010177612305,
709
- "total_bits": 449761824,
710
  "q_proj": {
711
  "group_size": {
712
  "2": 64
@@ -781,10 +781,10 @@
781
  },
782
  "down_proj": {
783
  "group_size": {
784
- "4": 128
785
  },
786
  "bits": [
787
- 4
788
  ],
789
  "bits_prop": [
790
  1
@@ -793,8 +793,8 @@
793
  }
794
  },
795
  "model.layers.9": {
796
- "accuracy": 0.9557886123657227,
797
- "total_bits": 360997440,
798
  "q_proj": {
799
  "group_size": {
800
  "2": 64
@@ -809,10 +809,10 @@
809
  },
810
  "k_proj": {
811
  "group_size": {
812
- "2": 64
813
  },
814
  "bits": [
815
- 2
816
  ],
817
  "bits_prop": [
818
  1
@@ -1057,8 +1057,8 @@
1057
  }
1058
  },
1059
  "model.layers.12": {
1060
- "accuracy": 0.9378215074539185,
1061
- "total_bits": 360997440,
1062
  "q_proj": {
1063
  "group_size": {
1064
  "2": 64
@@ -1073,10 +1073,10 @@
1073
  },
1074
  "k_proj": {
1075
  "group_size": {
1076
- "2": 64
1077
  },
1078
  "bits": [
1079
- 2
1080
  ],
1081
  "bits_prop": [
1082
  1
@@ -1233,8 +1233,8 @@
1233
  }
1234
  },
1235
  "model.layers.14": {
1236
- "accuracy": 0.9317333698272705,
1237
- "total_bits": 360997440,
1238
  "q_proj": {
1239
  "group_size": {
1240
  "2": 64
@@ -1249,10 +1249,10 @@
1249
  },
1250
  "k_proj": {
1251
  "group_size": {
1252
- "2": 64
1253
  },
1254
  "bits": [
1255
- 2
1256
  ],
1257
  "bits_prop": [
1258
  1
@@ -1761,8 +1761,8 @@
1761
  }
1762
  },
1763
  "model.layers.20": {
1764
- "accuracy": 0.9032342433929443,
1765
- "total_bits": 360997440,
1766
  "q_proj": {
1767
  "group_size": {
1768
  "2": 64
@@ -1777,10 +1777,10 @@
1777
  },
1778
  "k_proj": {
1779
  "group_size": {
1780
- "2": 64
1781
  },
1782
  "bits": [
1783
- 2
1784
  ],
1785
  "bits_prop": [
1786
  1
@@ -1849,8 +1849,8 @@
1849
  }
1850
  },
1851
  "model.layers.21": {
1852
- "accuracy": 0.9000318050384521,
1853
- "total_bits": 360997440,
1854
  "q_proj": {
1855
  "group_size": {
1856
  "2": 64
@@ -1865,10 +1865,10 @@
1865
  },
1866
  "k_proj": {
1867
  "group_size": {
1868
- "2": 64
1869
  },
1870
  "bits": [
1871
- 2
1872
  ],
1873
  "bits_prop": [
1874
  1
@@ -2289,8 +2289,8 @@
2289
  }
2290
  },
2291
  "model.layers.26": {
2292
- "accuracy": 0.9015021324157715,
2293
- "total_bits": 360997440,
2294
  "q_proj": {
2295
  "group_size": {
2296
  "2": 64
@@ -2305,10 +2305,10 @@
2305
  },
2306
  "k_proj": {
2307
  "group_size": {
2308
- "2": 64
2309
  },
2310
  "bits": [
2311
- 2
2312
  ],
2313
  "bits_prop": [
2314
  1
@@ -2465,8 +2465,8 @@
2465
  }
2466
  },
2467
  "model.layers.28": {
2468
- "accuracy": 0.9067087173461914,
2469
- "total_bits": 360997440,
2470
  "q_proj": {
2471
  "group_size": {
2472
  "2": 64
@@ -2541,10 +2541,10 @@
2541
  },
2542
  "down_proj": {
2543
  "group_size": {
2544
- "2": 64
2545
  },
2546
  "bits": [
2547
- 2
2548
  ],
2549
  "bits_prop": [
2550
  1
@@ -2553,8 +2553,8 @@
2553
  }
2554
  },
2555
  "model.layers.29": {
2556
- "accuracy": 0.9141254425048828,
2557
- "total_bits": 360997440,
2558
  "q_proj": {
2559
  "group_size": {
2560
  "2": 64
@@ -2629,10 +2629,10 @@
2629
  },
2630
  "down_proj": {
2631
  "group_size": {
2632
- "2": 64
2633
  },
2634
  "bits": [
2635
- 2
2636
  ],
2637
  "bits_prop": [
2638
  1
@@ -2641,8 +2641,8 @@
2641
  }
2642
  },
2643
  "model.layers.30": {
2644
- "accuracy": 0.9168109893798828,
2645
- "total_bits": 360997440,
2646
  "q_proj": {
2647
  "group_size": {
2648
  "2": 64
@@ -2717,10 +2717,10 @@
2717
  },
2718
  "down_proj": {
2719
  "group_size": {
2720
- "2": 64
2721
  },
2722
  "bits": [
2723
- 2
2724
  ],
2725
  "bits_prop": [
2726
  1
@@ -2729,8 +2729,8 @@
2729
  }
2730
  },
2731
  "model.layers.31": {
2732
- "accuracy": 0.9416450262069702,
2733
- "total_bits": 660324384,
2734
  "q_proj": {
2735
  "group_size": {
2736
  "2": 64
@@ -2745,10 +2745,10 @@
2745
  },
2746
  "k_proj": {
2747
  "group_size": {
2748
- "2": 64
2749
  },
2750
  "bits": [
2751
- 2
2752
  ],
2753
  "bits_prop": [
2754
  1
 
89
  }
90
  },
91
  "model.layers.1": {
92
+ "accuracy": 0.7773141860961914,
93
+ "total_bits": 365124672,
94
  "q_proj": {
95
  "group_size": {
96
  "2": 64
 
129
  },
130
  "o_proj": {
131
  "group_size": {
132
+ "2": 64
133
  },
134
  "bits": [
135
+ 2
136
  ],
137
  "bits_prop": [
138
  1
 
165
  },
166
  "down_proj": {
167
  "group_size": {
168
+ "2": 64
169
  },
170
  "bits": [
171
+ 2
172
  ],
173
  "bits_prop": [
174
  1
 
265
  }
266
  },
267
  "model.layers.3": {
268
+ "accuracy": 0.8372340202331543,
269
+ "total_bits": 360997440,
270
  "q_proj": {
271
  "group_size": {
272
  "2": 64
 
341
  },
342
  "down_proj": {
343
  "group_size": {
344
+ "2": 64
345
  },
346
  "bits": [
347
+ 2
348
  ],
349
  "bits_prop": [
350
  1
 
705
  }
706
  },
707
  "model.layers.8": {
708
+ "accuracy": 0.7149877548217773,
709
+ "total_bits": 360997440,
710
  "q_proj": {
711
  "group_size": {
712
  "2": 64
 
781
  },
782
  "down_proj": {
783
  "group_size": {
784
+ "2": 64
785
  },
786
  "bits": [
787
+ 2
788
  ],
789
  "bits_prop": [
790
  1
 
793
  }
794
  },
795
  "model.layers.9": {
796
+ "accuracy": 0.9577234387397766,
797
+ "total_bits": 365124672,
798
  "q_proj": {
799
  "group_size": {
800
  "2": 64
 
809
  },
810
  "k_proj": {
811
  "group_size": {
812
+ "4": 128
813
  },
814
  "bits": [
815
+ 4
816
  ],
817
  "bits_prop": [
818
  1
 
1057
  }
1058
  },
1059
  "model.layers.12": {
1060
+ "accuracy": 0.9404451847076416,
1061
+ "total_bits": 365124672,
1062
  "q_proj": {
1063
  "group_size": {
1064
  "2": 64
 
1073
  },
1074
  "k_proj": {
1075
  "group_size": {
1076
+ "4": 128
1077
  },
1078
  "bits": [
1079
+ 4
1080
  ],
1081
  "bits_prop": [
1082
  1
 
1233
  }
1234
  },
1235
  "model.layers.14": {
1236
+ "accuracy": 0.9359749555587769,
1237
+ "total_bits": 365124672,
1238
  "q_proj": {
1239
  "group_size": {
1240
  "2": 64
 
1249
  },
1250
  "k_proj": {
1251
  "group_size": {
1252
+ "4": 128
1253
  },
1254
  "bits": [
1255
+ 4
1256
  ],
1257
  "bits_prop": [
1258
  1
 
1761
  }
1762
  },
1763
  "model.layers.20": {
1764
+ "accuracy": 0.9059760570526123,
1765
+ "total_bits": 365124672,
1766
  "q_proj": {
1767
  "group_size": {
1768
  "2": 64
 
1777
  },
1778
  "k_proj": {
1779
  "group_size": {
1780
+ "4": 128
1781
  },
1782
  "bits": [
1783
+ 4
1784
  ],
1785
  "bits_prop": [
1786
  1
 
1849
  }
1850
  },
1851
  "model.layers.21": {
1852
+ "accuracy": 0.9023456573486328,
1853
+ "total_bits": 365124672,
1854
  "q_proj": {
1855
  "group_size": {
1856
  "2": 64
 
1865
  },
1866
  "k_proj": {
1867
  "group_size": {
1868
+ "4": 128
1869
  },
1870
  "bits": [
1871
+ 4
1872
  ],
1873
  "bits_prop": [
1874
  1
 
2289
  }
2290
  },
2291
  "model.layers.26": {
2292
+ "accuracy": 0.9034652709960938,
2293
+ "total_bits": 365124672,
2294
  "q_proj": {
2295
  "group_size": {
2296
  "2": 64
 
2305
  },
2306
  "k_proj": {
2307
  "group_size": {
2308
+ "4": 128
2309
  },
2310
  "bits": [
2311
+ 4
2312
  ],
2313
  "bits_prop": [
2314
  1
 
2465
  }
2466
  },
2467
  "model.layers.28": {
2468
+ "accuracy": 0.9271254539489746,
2469
+ "total_bits": 449761824,
2470
  "q_proj": {
2471
  "group_size": {
2472
  "2": 64
 
2541
  },
2542
  "down_proj": {
2543
  "group_size": {
2544
+ "4": 128
2545
  },
2546
  "bits": [
2547
+ 4
2548
  ],
2549
  "bits_prop": [
2550
  1
 
2553
  }
2554
  },
2555
  "model.layers.29": {
2556
+ "accuracy": 0.9337887763977051,
2557
+ "total_bits": 449761824,
2558
  "q_proj": {
2559
  "group_size": {
2560
  "2": 64
 
2629
  },
2630
  "down_proj": {
2631
  "group_size": {
2632
+ "4": 128
2633
  },
2634
  "bits": [
2635
+ 4
2636
  ],
2637
  "bits_prop": [
2638
  1
 
2641
  }
2642
  },
2643
  "model.layers.30": {
2644
+ "accuracy": 0.9370880126953125,
2645
+ "total_bits": 449761824,
2646
  "q_proj": {
2647
  "group_size": {
2648
  "2": 64
 
2717
  },
2718
  "down_proj": {
2719
  "group_size": {
2720
+ "4": 128
2721
  },
2722
  "bits": [
2723
+ 4
2724
  ],
2725
  "bits_prop": [
2726
  1
 
2729
  }
2730
  },
2731
  "model.layers.31": {
2732
+ "accuracy": 0.9427725076675415,
2733
+ "total_bits": 664451616,
2734
  "q_proj": {
2735
  "group_size": {
2736
  "2": 64
 
2745
  },
2746
  "k_proj": {
2747
  "group_size": {
2748
+ "4": 128
2749
  },
2750
  "bits": [
2751
+ 4
2752
  ],
2753
  "bits_prop": [
2754
  1
tokenizer.json CHANGED
@@ -14,20 +14,20 @@
14
  },
15
  {
16
  "id": 1,
17
- "content": "<s>",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
- "normalized": false,
22
  "special": true
23
  },
24
  {
25
  "id": 2,
26
- "content": "</s>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
30
- "normalized": false,
31
  "special": true
32
  },
33
  {
@@ -56,24 +56,6 @@
56
  "rstrip": false,
57
  "normalized": false,
58
  "special": true
59
- },
60
- {
61
- "id": 64000,
62
- "content": "<|startoftext|>",
63
- "single_word": false,
64
- "lstrip": false,
65
- "rstrip": false,
66
- "normalized": true,
67
- "special": true
68
- },
69
- {
70
- "id": 64001,
71
- "content": "<|endoftext|>",
72
- "single_word": false,
73
- "lstrip": false,
74
- "rstrip": false,
75
- "normalized": true,
76
- "special": true
77
  }
78
  ],
79
  "normalizer": {
@@ -153,8 +135,8 @@
153
  "byte_fallback": true,
154
  "vocab": {
155
  "<unk>": 0,
156
- "<s>": 1,
157
- "</s>": 2,
158
  "<|Human|>": 3,
159
  "<|Assistant|>": 4,
160
  "<|System|>": 5,
 
14
  },
15
  {
16
  "id": 1,
17
+ "content": "<|startoftext|>",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
21
+ "normalized": true,
22
  "special": true
23
  },
24
  {
25
  "id": 2,
26
+ "content": "<|endoftext|>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
30
+ "normalized": true,
31
  "special": true
32
  },
33
  {
 
56
  "rstrip": false,
57
  "normalized": false,
58
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
59
  }
60
  ],
61
  "normalizer": {
 
135
  "byte_fallback": true,
136
  "vocab": {
137
  "<unk>": 0,
138
+ "<|startoftext|>": 1,
139
+ "<|endoftext|>": 2,
140
  "<|Human|>": 3,
141
  "<|Assistant|>": 4,
142
  "<|System|>": 5,
tokenizer_config.json CHANGED
@@ -12,17 +12,17 @@
12
  "special": true
13
  },
14
  "1": {
15
- "content": "<s>",
16
  "lstrip": false,
17
- "normalized": false,
18
  "rstrip": false,
19
  "single_word": false,
20
  "special": true
21
  },
22
  "2": {
23
- "content": "</s>",
24
  "lstrip": false,
25
- "normalized": false,
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
@@ -50,22 +50,6 @@
50
  "rstrip": false,
51
  "single_word": false,
52
  "special": true
53
- },
54
- "64000": {
55
- "content": "<|startoftext|>",
56
- "lstrip": false,
57
- "normalized": true,
58
- "rstrip": false,
59
- "single_word": false,
60
- "special": true
61
- },
62
- "64001": {
63
- "content": "<|endoftext|>",
64
- "lstrip": false,
65
- "normalized": true,
66
- "rstrip": false,
67
- "single_word": false,
68
- "special": true
69
  }
70
  },
71
  "additional_special_tokens": [
 
12
  "special": true
13
  },
14
  "1": {
15
+ "content": "<|startoftext|>",
16
  "lstrip": false,
17
+ "normalized": true,
18
  "rstrip": false,
19
  "single_word": false,
20
  "special": true
21
  },
22
  "2": {
23
+ "content": "<|endoftext|>",
24
  "lstrip": false,
25
+ "normalized": true,
26
  "rstrip": false,
27
  "single_word": false,
28
  "special": true
 
50
  "rstrip": false,
51
  "single_word": false,
52
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  }
54
  },
55
  "additional_special_tokens": [