newsletter
commited on
Commit
•
77e8277
1
Parent(s):
794e033
Upload 3 files
Browse files
Codestral-22B-v0.1.Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:199147da3b36db1a45c09164fad7a1d3655e4fd3c4c60f58bdff61e9085cbee5
|
3 |
+
size 18252701888
|
Codestral-22B-v0.1.Q6_K.hash.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"kv_metadata": "236a353a2578864bd9ef988f68db598b6ddcd6bbe7225e706ec5639e811c54e1",
|
3 |
+
"tensor_metadata": "412eefc4bef7da4f94ff97ea87232aaa92b26cad41e43a410e731a2d9cc244b4",
|
4 |
+
"tensor_data": "9a701181596f7e5431755ad39492b0cd1778371aaa0bc7b77688d1d92a85c89a",
|
5 |
+
"metadata": "76cd885df98169b5e07d359dd632376edc3ea21bada00c0f1e7f87d412546ad1",
|
6 |
+
"overall": "b4a3e61256844818ea28a9bd6a43c13e16b73f4c0a646f6874eb9a446344cf6a"
|
7 |
+
}
|
Codestral-22B-v0.1.Q6_K.info.md
ADDED
@@ -0,0 +1,39 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Codestral-22B-v0.1.Q6_K.gguf - GGUF Internal File Dump
|
2 |
+
|
3 |
+
- Endian: LITTLE endian
|
4 |
+
|
5 |
+
## Key Value Metadata Store
|
6 |
+
|
7 |
+
There are 28 key-value pairs in this file
|
8 |
+
|
9 |
+
| POS | TYPE | Count | Key | Value |
|
10 |
+
|----:|:----------|------:|:---------------------------------------|:-----------------------------------------------------|
|
11 |
+
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
12 |
+
| 2 | UINT64 | 1 | GGUF.tensor_count | 507 |
|
13 |
+
| 3 | UINT64 | 1 | GGUF.kv_count | 25 |
|
14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
15 |
+
| 5 | STRING | 1 | general.name | `Codestral-22B-v0.1-hf-FIM-fix` |
|
16 |
+
| 6 | UINT32 | 1 | llama.block_count | 56 |
|
17 |
+
| 7 | UINT32 | 1 | llama.context_length | 32768 |
|
18 |
+
| 8 | UINT32 | 1 | llama.embedding_length | 6144 |
|
19 |
+
| 9 | UINT32 | 1 | llama.feed_forward_length | 16384 |
|
20 |
+
| 10 | UINT32 | 1 | llama.attention.head_count | 48 |
|
21 |
+
| 11 | UINT32 | 1 | llama.attention.head_count_kv | 8 |
|
22 |
+
| 12 | FLOAT32 | 1 | llama.rope.freq_base | 1000000.0 |
|
23 |
+
| 13 | FLOAT32 | 1 | llama.attention.layer_norm_rms_epsilon | 1e-05 |
|
24 |
+
| 14 | UINT32 | 1 | general.file_type | 18 |
|
25 |
+
| 15 | UINT32 | 1 | llama.vocab_size | 32768 |
|
26 |
+
| 16 | UINT32 | 1 | llama.rope.dimension_count | 128 |
|
27 |
+
| 17 | BOOL | 1 | tokenizer.ggml.add_space_prefix | False |
|
28 |
+
| 18 | STRING | 1 | tokenizer.ggml.model | `llama` |
|
29 |
+
| 19 | STRING | 1 | tokenizer.ggml.pre | `default` |
|
30 |
+
| 20 | [STRING] | 32768 | tokenizer.ggml.tokens | [ `<unk>`, `<s>`, `</s>`, `[INST]`, `[/INST]`, ... ] |
|
31 |
+
| 21 | [FLOAT32] | 32768 | tokenizer.ggml.scores | [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ... ] |
|
32 |
+
| 22 | [INT32] | 32768 | tokenizer.ggml.token_type | [ 2, 3, 3, 3, 3, 3, 3, ... ] |
|
33 |
+
| 23 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 1 |
|
34 |
+
| 24 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 2 |
|
35 |
+
| 25 | UINT32 | 1 | tokenizer.ggml.unknown_token_id | 0 |
|
36 |
+
| 26 | BOOL | 1 | tokenizer.ggml.add_bos_token | True |
|
37 |
+
| 27 | BOOL | 1 | tokenizer.ggml.add_eos_token | False |
|
38 |
+
| 28 | UINT32 | 1 | general.quantization_version | 2 |
|
39 |
+
|