update dumpfile with correct kv array rendering
Browse files
Tinyllama-4.6M-v0.0-F16.md → Tinyllama-4.6M-v0.0-F16.dump.md
RENAMED
@@ -11,20 +11,20 @@ There are 36 key-value pairs in this file
|
|
11 |
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
12 |
| 2 | UINT64 | 1 | GGUF.tensor_count | 75 |
|
13 |
| 3 | UINT64 | 1 | GGUF.kv_count | 33 |
|
14 |
-
| 4 | STRING | 1 | general.architecture |
|
15 |
-
| 5 | STRING | 1 | general.type |
|
16 |
-
| 6 | STRING | 1 | general.name |
|
17 |
-
| 7 | STRING | 1 | general.author |
|
18 |
-
| 8 | STRING | 1 | general.version |
|
19 |
-
| 9 | STRING | 1 | general.description |
|
20 |
-
| 10 | STRING | 1 | general.quantized_by |
|
21 |
-
| 11 | STRING | 1 | general.size_label |
|
22 |
-
| 12 | STRING | 1 | general.license |
|
23 |
-
| 13 | STRING | 1 | general.url |
|
24 |
-
| 14 | STRING | 1 | general.source.url |
|
25 |
-
| 15 | [STRING] | 5 | general.tags | [
|
26 |
-
| 16 | [STRING] | 1 | general.languages | [
|
27 |
-
| 17 | [STRING] | 2 | general.datasets | [
|
28 |
| 18 | UINT32 | 1 | llama.block_count | 8 |
|
29 |
| 19 | UINT32 | 1 | llama.context_length | 2048 |
|
30 |
| 20 | UINT32 | 1 | llama.embedding_length | 64 |
|
@@ -34,11 +34,11 @@ There are 36 key-value pairs in this file
|
|
34 |
| 24 | UINT32 | 1 | general.file_type | 1 |
|
35 |
| 25 | UINT32 | 1 | llama.vocab_size | 32000 |
|
36 |
| 26 | UINT32 | 1 | llama.rope.dimension_count | 4 |
|
37 |
-
| 27 | STRING | 1 | tokenizer.ggml.model |
|
38 |
-
| 28 | STRING | 1 | tokenizer.ggml.pre |
|
39 |
-
| 29 | [STRING] | 32000 | tokenizer.ggml.tokens | [
|
40 |
-
| 30 | [FLOAT32] | 32000 | tokenizer.ggml.scores | [
|
41 |
-
| 31 | [INT32] | 32000 | tokenizer.ggml.token_type | [
|
42 |
| 32 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 1 |
|
43 |
| 33 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 2 |
|
44 |
| 34 | UINT32 | 1 | tokenizer.ggml.unknown_token_id | 0 |
|
|
|
11 |
| 1 | UINT32 | 1 | GGUF.version | 3 |
|
12 |
| 2 | UINT64 | 1 | GGUF.tensor_count | 75 |
|
13 |
| 3 | UINT64 | 1 | GGUF.kv_count | 33 |
|
14 |
+
| 4 | STRING | 1 | general.architecture | `llama` |
|
15 |
+
| 5 | STRING | 1 | general.type | `model` |
|
16 |
+
| 6 | STRING | 1 | general.name | `TinyLLama` |
|
17 |
+
| 7 | STRING | 1 | general.author | `Maykeye` |
|
18 |
+
| 8 | STRING | 1 | general.version | `v0.0` |
|
19 |
+
| 9 | STRING | 1 | general.description | `This gguf is ported from a fir`...`M but using Llama architecture` |
|
20 |
+
| 10 | STRING | 1 | general.quantized_by | `Mofosyne` |
|
21 |
+
| 11 | STRING | 1 | general.size_label | `4.6M` |
|
22 |
+
| 12 | STRING | 1 | general.license | `apache-2.0` |
|
23 |
+
| 13 | STRING | 1 | general.url | `https://huggingface.co/mofosyne/TinyLLama-v0-llamafile` |
|
24 |
+
| 14 | STRING | 1 | general.source.url | `https://huggingface.co/Maykeye/TinyLLama-v0` |
|
25 |
+
| 15 | [STRING] | 5 | general.tags | [ `text generation`, `transformer`, `llama`, `tiny`, `tiny model` ] |
|
26 |
+
| 16 | [STRING] | 1 | general.languages | [ `en` ] |
|
27 |
+
| 17 | [STRING] | 2 | general.datasets | [ `https://hugging`...`-GPT4-train.txt`, `https://hugging`...`-GPT4-valid.txt` ] |
|
28 |
| 18 | UINT32 | 1 | llama.block_count | 8 |
|
29 |
| 19 | UINT32 | 1 | llama.context_length | 2048 |
|
30 |
| 20 | UINT32 | 1 | llama.embedding_length | 64 |
|
|
|
34 |
| 24 | UINT32 | 1 | general.file_type | 1 |
|
35 |
| 25 | UINT32 | 1 | llama.vocab_size | 32000 |
|
36 |
| 26 | UINT32 | 1 | llama.rope.dimension_count | 4 |
|
37 |
+
| 27 | STRING | 1 | tokenizer.ggml.model | `llama` |
|
38 |
+
| 28 | STRING | 1 | tokenizer.ggml.pre | `default` |
|
39 |
+
| 29 | [STRING] | 32000 | tokenizer.ggml.tokens | [ `<unk>`, `<s>`, `</s>`, `<0x00>`, `<0x01>`, ... ] |
|
40 |
+
| 30 | [FLOAT32] | 32000 | tokenizer.ggml.scores | [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ... ] |
|
41 |
+
| 31 | [INT32] | 32000 | tokenizer.ggml.token_type | [ 2, 3, 3, 6, 6, 6, 6, ... ] |
|
42 |
| 32 | UINT32 | 1 | tokenizer.ggml.bos_token_id | 1 |
|
43 |
| 33 | UINT32 | 1 | tokenizer.ggml.eos_token_id | 2 |
|
44 |
| 34 | UINT32 | 1 | tokenizer.ggml.unknown_token_id | 0 |
|
llama.cpp
CHANGED
@@ -1 +1 @@
|
|
1 |
-
Subproject commit
|
|
|
1 |
+
Subproject commit c3776cacabce2ee35f172fb72be7a519752125fa
|
llamafile-creation.sh
CHANGED
@@ -6,7 +6,7 @@ METADATA_FILE="maykeye_tinyllama-metadata.json"
|
|
6 |
###############################################################################
|
7 |
# Pull both model folder, llamafile (for the engine) and llama.cpp (for the conversion script)
|
8 |
echo == Prep Enviroment ==
|
9 |
-
git submodule update --init
|
10 |
|
11 |
###############################################################################
|
12 |
echo == Build and prep the llamafile engine execuable ==
|
@@ -37,6 +37,9 @@ echo == Convert from safetensor to gguf ==
|
|
37 |
./llama.cpp/convert_hf_to_gguf.py ${MODEL_DIR} --metadata ${METADATA_FILE} --outtype f16 --verbose
|
38 |
mv ${MODEL_DIR}/${OUTFILE}.gguf ${OUTFILE}.gguf
|
39 |
|
|
|
|
|
|
|
40 |
###############################################################################
|
41 |
echo == Generating Llamafile ==
|
42 |
cp ./llamafile/o/llama.cpp/main/main ${OUTFILE}.llamafile
|
@@ -53,7 +56,3 @@ EOF
|
|
53 |
###############################################################################
|
54 |
echo == Test Output ./${OUTFILE}.llamafile ==
|
55 |
./${OUTFILE}.llamafile --cli -p "hello world the gruff man said"
|
56 |
-
|
57 |
-
###############################################################################
|
58 |
-
echo == Useful GGUF Technical Dump ==
|
59 |
-
./llama.cpp/gguf-py/scripts/gguf_dump.py --markdown ${OUTFILE}.gguf > ${OUTFILE}.md
|
|
|
6 |
###############################################################################
|
7 |
# Pull both model folder, llamafile (for the engine) and llama.cpp (for the conversion script)
|
8 |
echo == Prep Enviroment ==
|
9 |
+
#git submodule update --init
|
10 |
|
11 |
###############################################################################
|
12 |
echo == Build and prep the llamafile engine execuable ==
|
|
|
37 |
./llama.cpp/convert_hf_to_gguf.py ${MODEL_DIR} --metadata ${METADATA_FILE} --outtype f16 --verbose
|
38 |
mv ${MODEL_DIR}/${OUTFILE}.gguf ${OUTFILE}.gguf
|
39 |
|
40 |
+
# Generate Diagnostics Dumpfile
|
41 |
+
./llama.cpp/gguf-py/scripts/gguf_dump.py --markdown ${OUTFILE}.gguf > ${OUTFILE}.dump.md
|
42 |
+
|
43 |
###############################################################################
|
44 |
echo == Generating Llamafile ==
|
45 |
cp ./llamafile/o/llama.cpp/main/main ${OUTFILE}.llamafile
|
|
|
56 |
###############################################################################
|
57 |
echo == Test Output ./${OUTFILE}.llamafile ==
|
58 |
./${OUTFILE}.llamafile --cli -p "hello world the gruff man said"
|
|
|
|
|
|
|
|