Text Generation
Transformers
GGUF
English
code
TensorBlock
GGUF
Eval Results
Inference Endpoints
morriszms commited on
Commit
d7d6acf
1 Parent(s): 3a06bcf

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Refact-1_6B-fim-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Refact-1_6B-fim-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Refact-1_6B-fim-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Refact-1_6B-fim-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ Refact-1_6B-fim-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ Refact-1_6B-fim-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ Refact-1_6B-fim-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ Refact-1_6B-fim-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ Refact-1_6B-fim-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ Refact-1_6B-fim-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ Refact-1_6B-fim-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ Refact-1_6B-fim-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,340 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ inference: true
4
+ widget:
5
+ - text: 'def print_hello_world():'
6
+ example_title: Hello world
7
+ group: Python
8
+ license: bigscience-openrail-m
9
+ pretrain-datasets:
10
+ - books
11
+ - arxiv
12
+ - c4
13
+ - falcon-refinedweb
14
+ - wiki
15
+ - github-issues
16
+ - stack_markdown
17
+ - self-made dataset of permissive github code
18
+ datasets:
19
+ - bigcode/the-stack-dedup
20
+ - rombodawg/2XUNCENSORED_MegaCodeTraining188k
21
+ - bigcode/commitpackft
22
+ metrics:
23
+ - code_eval
24
+ library_name: transformers
25
+ tags:
26
+ - code
27
+ - TensorBlock
28
+ - GGUF
29
+ language:
30
+ - en
31
+ base_model: smallcloudai/Refact-1_6B-fim
32
+ model-index:
33
+ - name: Refact-1.6B
34
+ results:
35
+ - task:
36
+ type: text-generation
37
+ dataset:
38
+ name: HumanEval
39
+ type: openai_humaneval
40
+ metrics:
41
+ - type: pass@1
42
+ value: 32.0
43
+ name: pass@1 (T=0.01)
44
+ verified: false
45
+ - type: pass@1
46
+ value: 31.5
47
+ name: pass@1 (T=0.2)
48
+ verified: false
49
+ - type: pass@10
50
+ value: 53.0
51
+ name: pass@10 (T=0.8)
52
+ verified: false
53
+ - type: pass@100
54
+ value: 76.9
55
+ name: pass@100 (T=0.8)
56
+ verified: false
57
+ - task:
58
+ type: text-generation
59
+ dataset:
60
+ name: HumanEvalSynthesize Python
61
+ type: bigcode/humanevalpack
62
+ metrics:
63
+ - type: pass@1
64
+ value: 35.8
65
+ name: pass@1 (T=0.2)
66
+ verified: false
67
+ - type: pass@1
68
+ value: 31.6
69
+ name: pass@1 (T=0.2)
70
+ verified: false
71
+ - type: pass@1
72
+ value: 29.1
73
+ name: pass@1 (T=0.2)
74
+ verified: false
75
+ - type: pass@1
76
+ value: -1
77
+ name: pass@1 (T=0.2)
78
+ verified: false
79
+ - type: pass@1
80
+ value: 26.3
81
+ name: pass@1 (T=0.2)
82
+ verified: false
83
+ - type: pass@1
84
+ value: -1
85
+ name: pass@1 (T=0.2)
86
+ verified: false
87
+ - type: pass@1
88
+ value: -1
89
+ name: pass@1 (T=0.2)
90
+ verified: false
91
+ - type: pass@1
92
+ value: 18.38
93
+ name: pass@1 (T=0.2)
94
+ verified: false
95
+ - type: pass@1
96
+ value: 12.28
97
+ name: pass@1 (T=0.2)
98
+ verified: false
99
+ - type: pass@1
100
+ value: 15.12
101
+ name: pass@1 (T=0.2)
102
+ verified: false
103
+ - type: pass@1
104
+ value: -1
105
+ name: pass@1 (T=0.2)
106
+ verified: false
107
+ - type: pass@1
108
+ value: 13.17
109
+ name: pass@1 (T=0.2)
110
+ verified: false
111
+ - type: pass@1
112
+ value: 2.8
113
+ name: pass@1 (T=0.2)
114
+ verified: false
115
+ - type: pass@1
116
+ value: -1
117
+ name: pass@1 (T=0.2)
118
+ verified: false
119
+ - type: pass@1
120
+ value: 26.92
121
+ name: pass@1 (T=0.2)
122
+ verified: false
123
+ - type: pass@1
124
+ value: 26.85
125
+ name: pass@1 (T=0.2)
126
+ verified: false
127
+ - type: pass@1
128
+ value: 30.76
129
+ name: pass@1 (T=0.2)
130
+ verified: false
131
+ - type: pass@1
132
+ value: -1
133
+ name: pass@1 (T=0.2)
134
+ verified: false
135
+ - type: pass@1
136
+ value: 25.94
137
+ name: pass@1 (T=0.2)
138
+ verified: false
139
+ - type: pass@1
140
+ value: 8.44
141
+ name: pass@1 (T=0.2)
142
+ verified: false
143
+ - type: pass@1
144
+ value: -1
145
+ name: pass@1 (T=0.2)
146
+ verified: false
147
+ - type: pass@1
148
+ value: 26.46
149
+ name: pass@1 (T=0.2)
150
+ verified: false
151
+ - type: pass@1
152
+ value: 17.86
153
+ name: pass@1 (T=0.2)
154
+ verified: false
155
+ - type: pass@1
156
+ value: 20.94
157
+ name: pass@1 (T=0.2)
158
+ verified: false
159
+ - type: pass@1
160
+ value: -1
161
+ name: pass@1 (T=0.2)
162
+ verified: false
163
+ - type: pass@1
164
+ value: 18.78
165
+ name: pass@1 (T=0.2)
166
+ verified: false
167
+ - type: pass@1
168
+ value: -1
169
+ name: pass@1 (T=0.2)
170
+ verified: false
171
+ - type: pass@1
172
+ value: -1
173
+ name: pass@1 (T=0.2)
174
+ verified: false
175
+ - task:
176
+ type: text-generation
177
+ dataset:
178
+ name: MBPP
179
+ type: mbpp
180
+ metrics:
181
+ - type: pass@1
182
+ value: 31.15
183
+ name: pass@1 (T=0.01)
184
+ verified: false
185
+ - task:
186
+ type: text-generation
187
+ dataset:
188
+ name: DS-1000 (Overall Completion)
189
+ type: ds1000
190
+ metrics:
191
+ - type: pass@1
192
+ value: 10.1
193
+ name: pass@1 (T=0.2)
194
+ verified: false
195
+ - task:
196
+ type: text-generation
197
+ dataset:
198
+ name: MultiPL-HumanEval (C++)
199
+ type: nuprl/MultiPL-E
200
+ metrics:
201
+ - type: pass@1
202
+ value: 21.61
203
+ name: pass@1 (T=0.2)
204
+ verified: false
205
+ - type: pass@1
206
+ value: 13.91
207
+ name: pass@1 (T=0.2)
208
+ verified: false
209
+ - type: pass@1
210
+ value: 9.5
211
+ name: pass@1 (T=0.2)
212
+ verified: false
213
+ - type: pass@1
214
+ value: 53.57
215
+ name: pass@1 (T=0.2)
216
+ verified: false
217
+ - type: pass@1
218
+ value: 21.58
219
+ name: pass@1 (T=0.2)
220
+ verified: false
221
+ - type: pass@1
222
+ value: 13.75
223
+ name: pass@1 (T=0.2)
224
+ verified: false
225
+ - type: pass@1
226
+ value: 26.88
227
+ name: pass@1 (T=0.2)
228
+ verified: false
229
+ - type: pass@1
230
+ value: 15.26
231
+ name: pass@1 (T=0.2)
232
+ verified: false
233
+ - type: pass@1
234
+ value: 23.04
235
+ name: pass@1 (T=0.2)
236
+ verified: false
237
+ - type: pass@1
238
+ value: 12.1
239
+ name: pass@1 (T=0.2)
240
+ verified: false
241
+ - type: pass@1
242
+ value: 29.6
243
+ name: pass@1 (T=0.2)
244
+ verified: false
245
+ - type: pass@1
246
+ value: 13.77
247
+ name: pass@1 (T=0.2)
248
+ verified: false
249
+ - type: pass@1
250
+ value: 12.68
251
+ name: pass@1 (T=0.2)
252
+ verified: false
253
+ - type: pass@1
254
+ value: 4.29
255
+ name: pass@1 (T=0.2)
256
+ verified: false
257
+ - type: pass@1
258
+ value: 19.54
259
+ name: pass@1 (T=0.2)
260
+ verified: false
261
+ - type: pass@1
262
+ value: 18.33
263
+ name: pass@1 (T=0.2)
264
+ verified: false
265
+ - type: pass@1
266
+ value: 5.7
267
+ name: pass@1 (T=0.2)
268
+ verified: false
269
+ - type: pass@1
270
+ value: 17.68
271
+ name: pass@1 (T=0.2)
272
+ verified: false
273
+ - type: pass@1
274
+ value: 25
275
+ name: pass@1 (T=0.2)
276
+ verified: false
277
+ ---
278
+
279
+ <div style="width: auto; margin-left: auto; margin-right: auto">
280
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
281
+ </div>
282
+ <div style="display: flex; justify-content: space-between; width: 100%;">
283
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
284
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
285
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
286
+ </p>
287
+ </div>
288
+ </div>
289
+
290
+ ## smallcloudai/Refact-1_6B-fim - GGUF
291
+
292
+ This repo contains GGUF format model files for [smallcloudai/Refact-1_6B-fim](https://huggingface.co/smallcloudai/Refact-1_6B-fim).
293
+
294
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
295
+
296
+ ## Prompt template
297
+
298
+ ```
299
+
300
+ ```
301
+
302
+ ## Model file specification
303
+
304
+ | Filename | Quant type | File Size | Description |
305
+ | -------- | ---------- | --------- | ----------- |
306
+ | [Refact-1_6B-fim-Q2_K.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q2_K.gguf) | Q2_K | 0.581 GB | smallest, significant quality loss - not recommended for most purposes |
307
+ | [Refact-1_6B-fim-Q3_K_S.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q3_K_S.gguf) | Q3_K_S | 0.673 GB | very small, high quality loss |
308
+ | [Refact-1_6B-fim-Q3_K_M.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q3_K_M.gguf) | Q3_K_M | 0.739 GB | very small, high quality loss |
309
+ | [Refact-1_6B-fim-Q3_K_L.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q3_K_L.gguf) | Q3_K_L | 0.795 GB | small, substantial quality loss |
310
+ | [Refact-1_6B-fim-Q4_0.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q4_0.gguf) | Q4_0 | 0.857 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
311
+ | [Refact-1_6B-fim-Q4_K_S.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q4_K_S.gguf) | Q4_K_S | 0.862 GB | small, greater quality loss |
312
+ | [Refact-1_6B-fim-Q4_K_M.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q4_K_M.gguf) | Q4_K_M | 0.902 GB | medium, balanced quality - recommended |
313
+ | [Refact-1_6B-fim-Q5_0.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q5_0.gguf) | Q5_0 | 1.030 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
314
+ | [Refact-1_6B-fim-Q5_K_S.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q5_K_S.gguf) | Q5_K_S | 1.030 GB | large, low quality loss - recommended |
315
+ | [Refact-1_6B-fim-Q5_K_M.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q5_K_M.gguf) | Q5_K_M | 1.053 GB | large, very low quality loss - recommended |
316
+ | [Refact-1_6B-fim-Q6_K.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q6_K.gguf) | Q6_K | 1.214 GB | very large, extremely low quality loss |
317
+ | [Refact-1_6B-fim-Q8_0.gguf](https://huggingface.co/tensorblock/Refact-1_6B-fim-GGUF/tree/main/Refact-1_6B-fim-Q8_0.gguf) | Q8_0 | 1.571 GB | very large, extremely low quality loss - not recommended |
318
+
319
+
320
+ ## Downloading instruction
321
+
322
+ ### Command line
323
+
324
+ Firstly, install Huggingface Client
325
+
326
+ ```shell
327
+ pip install -U "huggingface_hub[cli]"
328
+ ```
329
+
330
+ Then, downoad the individual model file the a local directory
331
+
332
+ ```shell
333
+ huggingface-cli download tensorblock/Refact-1_6B-fim-GGUF --include "Refact-1_6B-fim-Q2_K.gguf" --local-dir MY_LOCAL_DIR
334
+ ```
335
+
336
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
337
+
338
+ ```shell
339
+ huggingface-cli download tensorblock/Refact-1_6B-fim-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
340
+ ```
Refact-1_6B-fim-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4aef77b7d50fb404a7996b91d4c3e25d086acb5e72367c104faf2345d0fd321a
3
+ size 624292704
Refact-1_6B-fim-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74ed06cd2abfd37d3a2f28253aada6b386d0ed2f6f98271d06b72c685d2b1d99
3
+ size 853845856
Refact-1_6B-fim-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df450f57dd07aaa5db23bbdb646858a4fa419605db65463bebb41445105f5a95
3
+ size 793323360
Refact-1_6B-fim-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d1b5814004c90a291a1296cec4d7614b374807abd7a047a01a5da75a1c2d106
3
+ size 723003232
Refact-1_6B-fim-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a52894d28c8a623f784c0a36b925cd9b1b31b4442f23f97510f4376bb4b7ce7
3
+ size 920218464
Refact-1_6B-fim-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9c72df5570778f63a73db038be08c2715969e982519299d02d52a9ee87c5dc2
3
+ size 968338272
Refact-1_6B-fim-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f86b9224327d52022b3954f2b4f01fe497aecee0f3b1fb3595ba478d66e316e
3
+ size 926051168
Refact-1_6B-fim-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2e49cf8348f7560f44045f68e46acb1437b6aa9f5a0aed122bc874ab5ed0324
3
+ size 1105832800
Refact-1_6B-fim-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3cbd25896b50fd426f73dba45f1d7bdf2866ff7a88c063b7204532dafbf12c3
3
+ size 1130621792
Refact-1_6B-fim-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b33b7f1066d850dbdf4da92eabe3993f4753d7fae60c3212bdf8ae6356a82e8
3
+ size 1105832800
Refact-1_6B-fim-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e173a3a37aae820ba93261bf9e1279fbbbd8c5dbdecfb377aff277252ecda0b
3
+ size 1303048032
Refact-1_6B-fim-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37fe15a4fcf79257048316a0635b3a20e6c1b84930ca28944ad5041f2909ac51
3
+ size 1687086944