morriszms commited on
Commit
6d63a31
1 Parent(s): 0af8450

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ neural-chat-7b-v3-1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ neural-chat-7b-v3-1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ neural-chat-7b-v3-1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ neural-chat-7b-v3-1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ neural-chat-7b-v3-1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ neural-chat-7b-v3-1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ neural-chat-7b-v3-1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ neural-chat-7b-v3-1-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ neural-chat-7b-v3-1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ neural-chat-7b-v3-1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ neural-chat-7b-v3-1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ neural-chat-7b-v3-1-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - LLMs
5
+ - mistral
6
+ - Intel
7
+ - TensorBlock
8
+ - GGUF
9
+ pipeline_tag: text-generation
10
+ base_model: Intel/neural-chat-7b-v3-1
11
+ datasets:
12
+ - Open-Orca/SlimOrca
13
+ language:
14
+ - en
15
+ model-index:
16
+ - name: neural-chat-7b-v3-1
17
+ results:
18
+ - task:
19
+ type: Large Language Model
20
+ name: Large Language Model
21
+ dataset:
22
+ name: Open-Orca/SlimOrca
23
+ type: Open-Orca/SlimOrca
24
+ metrics:
25
+ - type: ARC (25-shot)
26
+ value: 66.21
27
+ name: ARC (25-shot)
28
+ verified: true
29
+ - type: HellaSwag (10-shot)
30
+ value: 83.64
31
+ name: HellaSwag (10-shot)
32
+ verified: true
33
+ - type: MMLU (5-shot)
34
+ value: 62.37
35
+ name: MMLU (5-shot)
36
+ verified: true
37
+ - type: TruthfulQA (0-shot)
38
+ value: 59.65
39
+ name: TruthfulQA (0-shot)
40
+ verified: true
41
+ - type: Winogrande (5-shot)
42
+ value: 78.14
43
+ name: Winogrande (5-shot)
44
+ verified: true
45
+ - type: GSM8K (5-shot)
46
+ value: 19.56
47
+ name: GSM8K (5-shot)
48
+ verified: true
49
+ - type: DROP (3-shot)
50
+ value: 43.84
51
+ name: DROP (3-shot)
52
+ verified: true
53
+ ---
54
+
55
+ <div style="width: auto; margin-left: auto; margin-right: auto">
56
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
57
+ </div>
58
+ <div style="display: flex; justify-content: space-between; width: 100%;">
59
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
60
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
61
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
62
+ </p>
63
+ </div>
64
+ </div>
65
+
66
+ ## Intel/neural-chat-7b-v3-1 - GGUF
67
+
68
+ This repo contains GGUF format model files for [Intel/neural-chat-7b-v3-1](https://huggingface.co/Intel/neural-chat-7b-v3-1).
69
+
70
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
71
+
72
+ ## Prompt template
73
+
74
+ ```
75
+ ### System:
76
+ {system_prompt}
77
+ ### User:
78
+ {prompt}
79
+ ### Assistant:
80
+ ```
81
+
82
+ ## Model file specification
83
+
84
+ | Filename | Quant type | File Size | Description |
85
+ | -------- | ---------- | --------- | ----------- |
86
+ | [neural-chat-7b-v3-1-Q2_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q2_K.gguf) | Q2_K | 2.532 GB | smallest, significant quality loss - not recommended for most purposes |
87
+ | [neural-chat-7b-v3-1-Q3_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q3_K_S.gguf) | Q3_K_S | 2.947 GB | very small, high quality loss |
88
+ | [neural-chat-7b-v3-1-Q3_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q3_K_M.gguf) | Q3_K_M | 3.277 GB | very small, high quality loss |
89
+ | [neural-chat-7b-v3-1-Q3_K_L.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q3_K_L.gguf) | Q3_K_L | 3.560 GB | small, substantial quality loss |
90
+ | [neural-chat-7b-v3-1-Q4_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q4_0.gguf) | Q4_0 | 3.827 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
91
+ | [neural-chat-7b-v3-1-Q4_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q4_K_S.gguf) | Q4_K_S | 3.856 GB | small, greater quality loss |
92
+ | [neural-chat-7b-v3-1-Q4_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q4_K_M.gguf) | Q4_K_M | 4.068 GB | medium, balanced quality - recommended |
93
+ | [neural-chat-7b-v3-1-Q5_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q5_0.gguf) | Q5_0 | 4.654 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
94
+ | [neural-chat-7b-v3-1-Q5_K_S.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q5_K_S.gguf) | Q5_K_S | 4.654 GB | large, low quality loss - recommended |
95
+ | [neural-chat-7b-v3-1-Q5_K_M.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q5_K_M.gguf) | Q5_K_M | 4.779 GB | large, very low quality loss - recommended |
96
+ | [neural-chat-7b-v3-1-Q6_K.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q6_K.gguf) | Q6_K | 5.534 GB | very large, extremely low quality loss |
97
+ | [neural-chat-7b-v3-1-Q8_0.gguf](https://huggingface.co/tensorblock/neural-chat-7b-v3-1-GGUF/tree/main/neural-chat-7b-v3-1-Q8_0.gguf) | Q8_0 | 7.167 GB | very large, extremely low quality loss - not recommended |
98
+
99
+
100
+ ## Downloading instruction
101
+
102
+ ### Command line
103
+
104
+ Firstly, install Huggingface Client
105
+
106
+ ```shell
107
+ pip install -U "huggingface_hub[cli]"
108
+ ```
109
+
110
+ Then, downoad the individual model file the a local directory
111
+
112
+ ```shell
113
+ huggingface-cli download tensorblock/neural-chat-7b-v3-1-GGUF --include "neural-chat-7b-v3-1-Q2_K.gguf" --local-dir MY_LOCAL_DIR
114
+ ```
115
+
116
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
117
+
118
+ ```shell
119
+ huggingface-cli download tensorblock/neural-chat-7b-v3-1-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
120
+ ```
neural-chat-7b-v3-1-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:069c39a8ac22a7164bbb5838d7d02fb371dc43fbfdf6243e00f73c7da122f6c7
3
+ size 2719243840
neural-chat-7b-v3-1-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0c79ef14f93f5e5b3e076f935888c1baf97e8a98e0688387286a1d301cf99f3
3
+ size 3822026304
neural-chat-7b-v3-1-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:442f7f1e19404f1ed4943e971027baed086131888e928b46a58ba586c9b6fa91
3
+ size 3518987840
neural-chat-7b-v3-1-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b47152e52431156c7fa2f14a9e33916bcf348145616db0d0f452ee00089efef7
3
+ size 3164569152
neural-chat-7b-v3-1-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a23b0918bb5a16ba34e00e128fd137e1d120527760ddca5e865fdf0a7baf8b49
3
+ size 4108918336
neural-chat-7b-v3-1-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba1c328b39809e5bd9568e64ee181541373023bb1c38c5d292028a8f2fdd4265
3
+ size 4368440896
neural-chat-7b-v3-1-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16db303c609e7c0d947bf2df1f10922c72859a90b291084c02f15fb940905bc3
3
+ size 4140375616
neural-chat-7b-v3-1-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7493949f7463f3bb9e1be0d9fdfa2f978b1df54f91bbdff4d368e6456ad9a7df
3
+ size 4997717568
neural-chat-7b-v3-1-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d886f14f054748292e3c5749cd9a77697152b7f21d1cbe0570492692d06e0014
3
+ size 5131411008
neural-chat-7b-v3-1-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:590d56d9d9ae0b5401b36b77d9be83f0efda3f7538d13f830a981c8afa6c26cc
3
+ size 4997717568
neural-chat-7b-v3-1-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c804a960ec38c8cbc2bb50f6688e81ca1420bf51589cc9004a13af5ad7d1a4d
3
+ size 5942066752
neural-chat-7b-v3-1-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6322ba3623f5966ad78222a69f84ef1d71ad3349433e4fe0fe5cf31933c105aa
3
+ size 7695859264