File size: 3,456 Bytes
dab9368
 
 
 
 
 
 
 
 
 
 
 
 
c64d94d
dab9368
23e5657
dab9368
20a2b96
 
87f327f
c64d94d
20a2b96
9d63fed
ac5c93c
20a2b96
dab9368
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
---
base_model:
- crestf411/L3-8B-sunfall-v0.4-stheno-v3.2
- Hastagaras/Jamet-8B-L3-MK.V-Blackroot
- Hastagaras/Jamet-8B-L3-MK.V-Blackroot
- TheDrummer/Llama-3SOME-8B-v2
tags:
- merge
- mergekit
- lazymergekit
- crestf411/L3-8B-sunfall-v0.4-stheno-v3.2
- Hastagaras/Jamet-8B-L3-MK.V-Blackroot
- TheDrummer/Llama-3SOME-8B-v2
- not-for-all-audiences
---
[GGUF Quants](https://huggingface.co/mradermacher/L3-Blackened-Sunfall-15B-GGUF?not-for-all-audiences=true) by [mradermacher](https://huggingface.co/mradermacher) 

[iGGUF Quants](https://huggingface.co/mradermacher/L3-Blackened-Sunfall-15B-i1-GGUF) by [mradermacher](https://huggingface.co/mradermacher) 

My first merge and I honestly thought it would go worse, but I'm pleased with the results. Originally, this model was created to be merged with another model as a part of a personal project; now known as [Blackfall Summanus](https://huggingface.co/kromeurus/L3-Blackfall-Summanus-v0.1-15B).
I wanted a 15B eRP model with the 'humanness' of Jamet-Blackroot with the chaos of Sunfall Stheno, tamed a little by some 'lewd' per se. It worked, mostly. By all means, this is an experimental merge, I'm sure there are problems I haven't found so please bear with me.

In limited testing, this model can't handle high temps at all. Anything higher then 1.3 will send it off the rails. Seems to play better with top P=0.8 and top K=25 on Llama3 template, but I'm still fooling with the settings. Has a tendency to ramble and can get quite dramatic, nothing too insane yet.
Interently a little buggy and incoherent so proceed with caution. 

# L3-Blackened-Sunfall-15B

L3-Blackened-Sunfall-15B is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [crestf411/L3-8B-sunfall-v0.4-stheno-v3.2](https://huggingface.co/crestf411/L3-8B-sunfall-v0.4-stheno-v3.2)
* [Hastagaras/Jamet-8B-L3-MK.V-Blackroot](https://huggingface.co/Hastagaras/Jamet-8B-L3-MK.V-Blackroot)
* [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2)

## 🧩 Configuration

```yaml
dtype: bfloat16
merge_method: passthrough
slices:
- sources:
  - layer_range: [0, 20]
    model: crestf411/L3-8B-sunfall-v0.4-stheno-v3.2
- sources:
  - layer_range: [6, 24]
    model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
    parameters:
      scale:
      - filter: o_proj
        value: 0.0
      - filter: down_proj
        value: 0.0
      - value: 1.0
- sources:
  - layer_range: [8, 26]
    model: Hastagaras/Jamet-8B-L3-MK.V-Blackroot
    parameters:
      scale:
      - filter: o_proj
        value: 0.0
      - filter: down_proj
        value: 0.0
      - value: 1.0
- sources:
  - layer_range: [24, 32]
    model: TheDrummer/Llama-3SOME-8B-v2
```

## 💻 Usage

```python
!pip install -qU transformers accelerate

from transformers import AutoTokenizer
import transformers
import torch

model = "kromeurus/L3-Blackened-Sunfall-15B"
messages = [{"role": "user", "content": "What is a large language model?"}]

tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
    "text-generation",
    model=model,
    torch_dtype=torch.float16,
    device_map="auto",
)

outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
```