morriszms commited on
Commit
de82ed2
·
verified ·
1 Parent(s): e290890

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,15 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ gemma-2-9b-it-tr-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ gemma-2-9b-it-tr-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
38
+ gemma-2-9b-it-tr-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ gemma-2-9b-it-tr-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
40
+ gemma-2-9b-it-tr-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
41
+ gemma-2-9b-it-tr-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
42
+ gemma-2-9b-it-tr-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
43
+ gemma-2-9b-it-tr-Q5_0.gguf filter=lfs diff=lfs merge=lfs -text
44
+ gemma-2-9b-it-tr-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
45
+ gemma-2-9b-it-tr-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
46
+ gemma-2-9b-it-tr-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
47
+ gemma-2-9b-it-tr-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,115 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: gemma
4
+ language:
5
+ - tr
6
+ base_model: neuralwork/gemma-2-9b-it-tr
7
+ pipeline_tag: text-generation
8
+ tags:
9
+ - TensorBlock
10
+ - GGUF
11
+ model-index:
12
+ - name: neuralwork/gemma-2-9b-it-tr
13
+ results:
14
+ - task:
15
+ type: multiple-choice
16
+ dataset:
17
+ name: MMLU_TR_V0.2
18
+ type: multiple-choice
19
+ metrics:
20
+ - type: 5-shot
21
+ value: 0.6117
22
+ name: 5-shot
23
+ verified: true
24
+ - type: 0-shot
25
+ value: 0.5583
26
+ name: 0-shot
27
+ verified: true
28
+ - type: 25-shot
29
+ value: 0.564
30
+ name: 25-shot
31
+ verified: true
32
+ - type: 10-shot
33
+ value: 0.5646
34
+ name: 10-shot
35
+ verified: true
36
+ - type: 5-shot
37
+ value: 0.6211
38
+ name: 5-shot
39
+ verified: true
40
+ - type: 5-shot
41
+ value: 0.6209
42
+ name: 5-shot
43
+ verified: true
44
+ ---
45
+
46
+ <div style="width: auto; margin-left: auto; margin-right: auto">
47
+ <img src="https://i.imgur.com/jC7kdl8.jpeg" alt="TensorBlock" style="width: 100%; min-width: 400px; display: block; margin: auto;">
48
+ </div>
49
+ <div style="display: flex; justify-content: space-between; width: 100%;">
50
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
51
+ <p style="margin-top: 0.5em; margin-bottom: 0em;">
52
+ Feedback and support: TensorBlock's <a href="https://x.com/tensorblock_aoi">Twitter/X</a>, <a href="https://t.me/TensorBlock">Telegram Group</a> and <a href="https://x.com/tensorblock_aoi">Discord server</a>
53
+ </p>
54
+ </div>
55
+ </div>
56
+
57
+ ## neuralwork/gemma-2-9b-it-tr - GGUF
58
+
59
+ This repo contains GGUF format model files for [neuralwork/gemma-2-9b-it-tr](https://huggingface.co/neuralwork/gemma-2-9b-it-tr).
60
+
61
+ The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4823](https://github.com/ggml-org/llama.cpp/commit/5bbe6a9fe9a8796a9389c85accec89dbc4d91e39).
62
+
63
+ <div style="text-align: left; margin: 20px 0;">
64
+ <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
65
+ Run them on the TensorBlock client using your local machine ↗
66
+ </a>
67
+ </div>
68
+
69
+ ## Prompt template
70
+
71
+ ```
72
+ <bos><start_of_turn>user
73
+ {prompt}<end_of_turn>
74
+ <start_of_turn>model
75
+ ```
76
+
77
+ ## Model file specification
78
+
79
+ | Filename | Quant type | File Size | Description |
80
+ | -------- | ---------- | --------- | ----------- |
81
+ | [gemma-2-9b-it-tr-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q2_K.gguf) | Q2_K | 3.805 GB | smallest, significant quality loss - not recommended for most purposes |
82
+ | [gemma-2-9b-it-tr-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q3_K_S.gguf) | Q3_K_S | 4.338 GB | very small, high quality loss |
83
+ | [gemma-2-9b-it-tr-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q3_K_M.gguf) | Q3_K_M | 4.762 GB | very small, high quality loss |
84
+ | [gemma-2-9b-it-tr-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q3_K_L.gguf) | Q3_K_L | 5.132 GB | small, substantial quality loss |
85
+ | [gemma-2-9b-it-tr-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q4_0.gguf) | Q4_0 | 5.443 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
86
+ | [gemma-2-9b-it-tr-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q4_K_S.gguf) | Q4_K_S | 5.479 GB | small, greater quality loss |
87
+ | [gemma-2-9b-it-tr-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q4_K_M.gguf) | Q4_K_M | 5.761 GB | medium, balanced quality - recommended |
88
+ | [gemma-2-9b-it-tr-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q5_0.gguf) | Q5_0 | 6.484 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
89
+ | [gemma-2-9b-it-tr-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q5_K_S.gguf) | Q5_K_S | 6.484 GB | large, low quality loss - recommended |
90
+ | [gemma-2-9b-it-tr-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q5_K_M.gguf) | Q5_K_M | 6.647 GB | large, very low quality loss - recommended |
91
+ | [gemma-2-9b-it-tr-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q6_K.gguf) | Q6_K | 7.589 GB | very large, extremely low quality loss |
92
+ | [gemma-2-9b-it-tr-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-9b-it-tr-GGUF/blob/main/gemma-2-9b-it-tr-Q8_0.gguf) | Q8_0 | 9.827 GB | very large, extremely low quality loss - not recommended |
93
+
94
+
95
+ ## Downloading instruction
96
+
97
+ ### Command line
98
+
99
+ Firstly, install Huggingface Client
100
+
101
+ ```shell
102
+ pip install -U "huggingface_hub[cli]"
103
+ ```
104
+
105
+ Then, downoad the individual model file the a local directory
106
+
107
+ ```shell
108
+ huggingface-cli download tensorblock/gemma-2-9b-it-tr-GGUF --include "gemma-2-9b-it-tr-Q2_K.gguf" --local-dir MY_LOCAL_DIR
109
+ ```
110
+
111
+ If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf`), you can try:
112
+
113
+ ```shell
114
+ huggingface-cli download tensorblock/gemma-2-9b-it-tr-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
115
+ ```
gemma-2-9b-it-tr-Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36dd2685126140a7418fe398154625a13ff8ed7511c60c0cedb446bfc72cc9b7
3
+ size 3805398272
gemma-2-9b-it-tr-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bbebd323a26e4a8ddade3198e31e3ee906d320cacfafa1a795efc38ed113319
3
+ size 5132453120
gemma-2-9b-it-tr-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b085022fb3e7b1b3155ec6fc5cee0ece73f9b3832d3a5b2ce30f2ccd19773c9c
3
+ size 4761781504
gemma-2-9b-it-tr-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:502b831f57da1344bf2a12746cb0aab94341d248abeada6d99d63169c6203257
3
+ size 4337665280
gemma-2-9b-it-tr-Q4_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2caf01b11707cc3038c8192d1e571627617a3893ed5966da77e37b318ad2c5f
3
+ size 5443142912
gemma-2-9b-it-tr-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:721bab2440ced40fafa8708e3b3b132df610dfcf23e3eff229ece050ab498a4f
3
+ size 5761058048
gemma-2-9b-it-tr-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cbb7f2f7b9280f8df7b55fd31099d239b9c4c85336a6de70dc2131585074c33
3
+ size 5478925568
gemma-2-9b-it-tr-Q5_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91c737af4789adf3997b51389369db4117295487e5ad11452dd838e83df976da
3
+ size 6483592448
gemma-2-9b-it-tr-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:003b7a759f71c4966fe73cadabdacea49a853bd9d89cb4cc1921b37ce99c3524
3
+ size 6647366912
gemma-2-9b-it-tr-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b16da45f2d0ed81c5c60fade088876cc38d2d19fe10f54fa30c3952f38fa03a
3
+ size 6483592448
gemma-2-9b-it-tr-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c101b44977a27451c71016012951eddbfe684900abdb46dd01b97a2aab9f7c4b
3
+ size 7589070080
gemma-2-9b-it-tr-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0dac0722e6f1f4e9eceda2176d73b2da8081c1ae995bbf519a4a196fe88908e2
3
+ size 9827149056