grimjim commited on
Commit
99dbcd2
1 Parent(s): bdc8bef

Update README.md

Browse files

Updated with links to GGUF and fp16.

Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -12,7 +12,11 @@ license: cc-by-nc-4.0
12
 
13
  This is an 8.0bpw h8 exl2 quant of a merger of two similar models with strong reasoning, hopefully resulting in "dense" encoding of said reasoning, was merged with a model targeting roleplay.
14
 
15
- I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I've only found it stable up to 8K in testing. I recommend sticking with 8.0bpw h8 exl2 or Q8_0 GGUF, to maintain coherence.
 
 
 
 
16
 
17
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
18
 
 
12
 
13
  This is an 8.0bpw h8 exl2 quant of a merger of two similar models with strong reasoning, hopefully resulting in "dense" encoding of said reasoning, was merged with a model targeting roleplay.
14
 
15
+ I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I found it lost coherence after 8K in informal testing. I prefer to stick with 8.0bpw h8 exl2 or Q8_0 GGUF for maximum coherence.
16
+
17
+ Alternative downloads:
18
+ [GGUF quants courtesy of Lewdiculous](https://huggingface.co/Lewdiculous/kukulemon-7B-GGUF-IQ-Imatrix)
19
+ [fp16 safetensors](https://huggingface.co/grimjim/kukulemon-7B)
20
 
21
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
22