Update README.md
Browse filesUpdated description. Minor corrections.
README.md
CHANGED
@@ -6,9 +6,13 @@ library_name: transformers
|
|
6 |
tags:
|
7 |
- mergekit
|
8 |
- merge
|
9 |
-
|
10 |
---
|
11 |
-
#
|
|
|
|
|
|
|
|
|
12 |
|
13 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
14 |
|
@@ -47,4 +51,4 @@ parameters:
|
|
47 |
- value: 0.5 # fallback for rest of tensors
|
48 |
dtype: float16
|
49 |
|
50 |
-
```
|
|
|
6 |
tags:
|
7 |
- mergekit
|
8 |
- merge
|
9 |
+
license: cc-by-nc-4.0
|
10 |
---
|
11 |
+
# kukulemon-7B
|
12 |
+
|
13 |
+
A merger of two similar models with strong reasoning, hopefully resulting in "dense" encoding of said reasoning, was merged with a model targeting roleplay.
|
14 |
+
|
15 |
+
I've tested with ChatML prompts with temperature=1.1 and minP=0.03. The model itself supports Alpaca format prompts. The model claims a context length of 32K, but I've only tested to 8K to date.
|
16 |
|
17 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
18 |
|
|
|
51 |
- value: 0.5 # fallback for rest of tensors
|
52 |
dtype: float16
|
53 |
|
54 |
+
```
|