grimjim commited on
Commit
2bc260c
1 Parent(s): 83056ed

Update README.md

Browse files

Updated links to quants

Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -7,6 +7,7 @@ tags:
7
  - mergekit
8
  - merge
9
  license: cc-by-nc-4.0
 
10
  ---
11
  # kukulemon-7B-8.0bpw-h8_exl2
12
 
@@ -15,9 +16,9 @@ This is an 8.0bpw h8 exl2 quant of a merger of two similar models with strong re
15
  I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I found it lost coherence after 8K in informal testing. I prefer to stick with 8.0bpw h8 exl2 or Q8_0 GGUF for maximum coherence.
16
 
17
  Alternative downloads:
18
- [GGUF quants courtesy of Lewdiculous](https://huggingface.co/Lewdiculous/kukulemon-7B-GGUF-IQ-Imatrix)
19
- [fp16 safetensors](https://huggingface.co/grimjim/kukulemon-7B)
20
- [Q8_0 GGUF](https://huggingface.co/grimjim/kukulemon-7B-GGUF)
21
 
22
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
23
 
 
7
  - mergekit
8
  - merge
9
  license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
  ---
12
  # kukulemon-7B-8.0bpw-h8_exl2
13
 
 
16
  I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I found it lost coherence after 8K in informal testing. I prefer to stick with 8.0bpw h8 exl2 or Q8_0 GGUF for maximum coherence.
17
 
18
  Alternative downloads:
19
+ - [iMatrix GGUF quants courtesy of Lewdiculous](https://huggingface.co/Lewdiculous/kukulemon-7B-GGUF-IQ-Imatrix)
20
+ - [fp16 safetensors](https://huggingface.co/grimjim/kukulemon-7B)
21
+ - [GGUF quants](https://huggingface.co/grimjim/kukulemon-7B-GGUF)
22
 
23
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
24