morriszms commited on
Commit
eecec1b
·
verified ·
1 Parent(s): a2a8db7

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ neural-chat-7b-v3-3-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ neural-chat-7b-v3-3-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ neural-chat-7b-v3-3-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ neural-chat-7b-v3-3-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ neural-chat-7b-v3-3-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ neural-chat-7b-v3-3-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ neural-chat-7b-v3-3-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ neural-chat-7b-v3-3-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ neural-chat-7b-v3-3-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ neural-chat-7b-v3-3-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ neural-chat-7b-v3-3-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ neural-chat-7b-v3-3-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,212 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - LLMs
5
+ - mistral
6
+ - math
7
+ - Intel
8
+ - TensorBlock
9
+ - GGUF
10
+ base_model: Intel/neural-chat-7b-v3-3
11
+ model-index:
12
+ - name: neural-chat-7b-v3-3
13
+ results:
14
+ - task:
15
+ type: Large Language Model
16
+ name: Large Language Model
17
+ dataset:
18
+ name: meta-math/MetaMathQA
19
+ type: meta-math/MetaMathQA
20
+ metrics:
21
+ - type: ARC (25-shot)
22
+ value: 66.89
23
+ name: ARC (25-shot)
24
+ verified: true
25
+ - type: HellaSwag (10-shot)
26
+ value: 85.26
27
+ name: HellaSwag (10-shot)
28
+ verified: true
29
+ - type: MMLU (5-shot)
30
+ value: 63.07
31
+ name: MMLU (5-shot)
32
+ verified: true
33
+ - type: TruthfulQA (0-shot)
34
+ value: 63.01
35
+ name: TruthfulQA (0-shot)
36
+ verified: true
37
+ - type: Winogrande (5-shot)
38
+ value: 79.64
39
+ name: Winogrande (5-shot)
40
+ verified: true
41
+ - type: GSM8K (5-shot)
42
+ value: 61.11
43
+ name: GSM8K (5-shot)
44
+ verified: true
45
+ - task:
46
+ type: text-generation
47
+ name: Text Generation
48
+ dataset:
49
+ name: AI2 Reasoning Challenge (25-Shot)
50
+ type: ai2_arc
51
+ config: ARC-Challenge
52
+ split: test
53
+ args:
54
+ num_few_shot: 25
55
+ metrics:
56
+ - type: acc_norm
57
+ value: 66.89
58
+ name: normalized accuracy
59
+ source:
60
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
61
+ name: Open LLM Leaderboard
62
+ - task:
63
+ type: text-generation
64
+ name: Text Generation
65
+ dataset:
66
+ name: HellaSwag (10-Shot)
67
+ type: hellaswag
68
+ split: validation
69
+ args:
70
+ num_few_shot: 10
71
+ metrics:
72
+ - type: acc_norm
73
+ value: 85.26
74
+ name: normalized accuracy
75
+ source:
76
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
77
+ name: Open LLM Leaderboard
78
+ - task:
79
+ type: text-generation
80
+ name: Text Generation
81
+ dataset:
82
+ name: MMLU (5-Shot)
83
+ type: cais/mmlu
84
+ config: all
85
+ split: test
86
+ args:
87
+ num_few_shot: 5
88
+ metrics:
89
+ - type: acc
90
+ value: 63.07
91
+ name: accuracy
92
+ source:
93
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
94
+ name: Open LLM Leaderboard
95
+ - task:
96
+ type: text-generation
97
+ name: Text Generation
98
+ dataset:
99
+ name: TruthfulQA (0-shot)
100
+ type: truthful_qa
101
+ config: multiple_choice
102
+ split: validation
103
+ args:
104
+ num_few_shot: 0
105
+ metrics:
106
+ - type: mc2
107
+ value: 63.01
108
+ source:
109
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
110
+ name: Open LLM Leaderboard
111
+ - task:
112
+ type: text-generation
113
+ name: Text Generation
114
+ dataset:
115
+ name: Winogrande (5-shot)
116
+ type: winogrande
117
+ config: winogrande_xl
118
+ split: validation
119
+ args:
120
+ num_few_shot: 5
121
+ metrics:
122
+ - type: acc
123
+ value: 79.64
124
+ name: accuracy
125
+ source:
126
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
127
+ name: Open LLM Leaderboard
128
+ - task:
129
+ type: text-generation
130
+ name: Text Generation
131
+ dataset:
132
+ name: GSM8k (5-shot)
133
+ type: gsm8k
134
+ config: main
135
+ split: test
136
+ args:
137
+ num_few_shot: 5
138
+ metrics:
139
+ - type: acc
140
+ value: 61.11
141
+ name: accuracy
142
+ source:
143
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Intel/neural-chat-7b-v3-3
144
+ name: Open LLM Leaderboard
145
+ ---
146
+
147
+ <div style="width: auto; margin-left: auto; margin-right: auto">
148
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
149
+ </div>
150
+ <div style="display: flex; justify-content: space-between; width: 100%;">
151
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
152
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
153
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
154
+ </p>
155
+ </div>
156
+ </div>
157
+
158
+ ## Intel/neural-chat-7b-v3-3 - GGUF
159
+
160
+ This repo contains GGUF format model files for [Intel/neural-chat-7b-v3-3](https://huggingface.co/Intel/neural-chat-7b-v3-3).
161
+
162
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
163
+
164
+ ## Prompt template
165
+
166
+ ```
167
+ ### System:
168
+ {system_prompt}
169
+ ### User:
170
+ {prompt}
171
+ ### Assistant:
172
+ ```
173
+
174
+ ## Model file specification
175
+
176
+ | Filename | Quant type | File Size | Description |
177
+ | -------- | ---------- | --------- | ----------- |
178
+ | [neural-chat-7b-v3-3-Q2_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q2_K.gguf) | Q2_K | 2.532 GB | smallest, significant quality loss - not recommended for most purposes |
179
+ | [neural-chat-7b-v3-3-Q3_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q3_K_S.gguf) | Q3_K_S | 2.947 GB | very small, high quality loss |
180
+ | [neural-chat-7b-v3-3-Q3_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q3_K_M.gguf) | Q3_K_M | 3.277 GB | very small, high quality loss |
181
+ | [neural-chat-7b-v3-3-Q3_K_L.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q3_K_L.gguf) | Q3_K_L | 3.560 GB | small, substantial quality loss |
182
+ | [neural-chat-7b-v3-3-Q4_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q4_0.gguf) | Q4_0 | 3.827 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
183
+ | [neural-chat-7b-v3-3-Q4_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q4_K_S.gguf) | Q4_K_S | 3.856 GB | small, greater quality loss |
184
+ | [neural-chat-7b-v3-3-Q4_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q4_K_M.gguf) | Q4_K_M | 4.068 GB | medium, balanced quality - recommended |
185
+ | [neural-chat-7b-v3-3-Q5_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q5_0.gguf) | Q5_0 | 4.654 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
186
+ | [neural-chat-7b-v3-3-Q5_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q5_K_S.gguf) | Q5_K_S | 4.654 GB | large, low quality loss - recommended |
187
+ | [neural-chat-7b-v3-3-Q5_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q5_K_M.gguf) | Q5_K_M | 4.779 GB | large, very low quality loss - recommended |
188
+ | [neural-chat-7b-v3-3-Q6_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q6_K.gguf) | Q6_K | 5.534 GB | very large, extremely low quality loss |
189
+ | [neural-chat-7b-v3-3-Q8_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-3-GGUF/tree/main/neural-chat-7b-v3-3-Q8_0.gguf) | Q8_0 | 7.167 GB | very large, extremely low quality loss - not recommended |
190
+
191
+
192
+ ## Downloading instruction
193
+
194
+ ### Command line
195
+
196
+ Firstly, install Huggingface Client
197
+
198
+ ```shell
199
+ pip install -U "huggingface_hub[cli]"
200
+ ```
201
+
202
+ Then, downoad the individual model file the a local directory
203
+
204
+ ```shell
205
+ huggingface-cli download tensorblock/neural-chat-7b-v3-3-GGUF --include "neural-chat-7b-v3-3-Q2_K.gguf" --local-dir MY_LOCAL_DIR
206
+ ```
207
+
208
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
209
+
210
+ ```shell
211
+ huggingface-cli download tensorblock/neural-chat-7b-v3-3-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
212
+ ```
neural-chat-7b-v3-3-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11e61d31a419ff6dd4d035bc08fd72b1704d0b3a0e5814dd1df4a5df083b93cb
3
+ size 2719243680
neural-chat-7b-v3-3-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5892c6d580c624720c428cfd5704667da786183631a8de0fce89e7ab8d8f4760
3
+ size 3822026144
neural-chat-7b-v3-3-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b938f550c0265d2147fb4a871e0ff457268e58ceef63951aea536d5cb82ec677
3
+ size 3518987680
neural-chat-7b-v3-3-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f137edf06bebe4d5fffbf3984c75eb17145f6d63f5699f97b6048f5716d4b96a
3
+ size 3164568992
neural-chat-7b-v3-3-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35c20212ddf185796a2f677fd06234417f790d99108762dd43f0ed5a5d9ef504
3
+ size 4108918176
neural-chat-7b-v3-3-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:372a10cabc30b7a5cb85c2ab8fb1a08741fe8d7f3784d508591f1eeb078b9c20
3
+ size 4368440736
neural-chat-7b-v3-3-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:447d009894f3ba4e42dc92ae185f35d6eab8ea4701a10cf0ad0d0bab655591bd
3
+ size 4140375456
neural-chat-7b-v3-3-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a35349bdebf24c5d93caa9b19ea460f45825871c2caf284cc36f5d89133934a1
3
+ size 4997717408
neural-chat-7b-v3-3-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e291ee602066d127f066f78cc813c410839957181fb6377884084e22d3264fc8
3
+ size 5131410848
neural-chat-7b-v3-3-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:320b0e0672422d751c56d3cd756c3cd1dc8d3056ba46b78525f36674bee437aa
3
+ size 4997717408
neural-chat-7b-v3-3-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c71322a3a4e26c1605d82c363531cad17fc10cd4dbf42c7a1e41b65b60125657
3
+ size 5942066592
neural-chat-7b-v3-3-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:752c83ae008629dad42d48b5e968b9d0e9c3a7accc1de04474f7e5e344f1375b
3
+ size 7695859104