File size: 2,266 Bytes
39f2e99
54f11b9
 
 
 
 
 
39f2e99
54f11b9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
---

base_model: []
library_name: transformers
tags:
- mergekit
- merge

---

# model
Llama-3-70b-Arimas-story-RP-V1.6

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
I Greatly expanded the amount of models used in this merge, experimented a lot with different idea's.
This version feels a lot more convincing than V1.5 Hopefully the long context window will also remain strong after Quants.
Because of the many merges switched back from BFloat to Float.
Tried breadcrums without the Ties, that went very poorly.

### Merge Method

This model was merged using the breadcrumbs_ties merge method using I:\Llama-3-70B-Instruct-Gradient-262k as a base.



### Models Merged



The following models were included in the merge:

* \Smaug-Llama-3-70B-Instruct

* \Meta-LLama-3-Cat-Smaug-LLama-70b

* \Meta-LLama-3-Cat-A-LLama-70b

* \Llama-3-70B-Synthia-v3.5

* \Llama-3-70B-Instruct-Gradient-524k

* \Llama-3-70B-Instruct-Gradient-262k

* \Tess-2.0-Llama-3-70B-v0.2

* \Llama-3-Lumimaid-70B-v0.1-alt



### Configuration



The following YAML configuration was used to produce this model:



```yaml

models:

  - model: \Llama-3-70B-Instruct-Gradient-262k

    parameters:

      weight: 0.25

      density: 0.90

      gamma: 0.01

  - model: \Meta-LLama-3-Cat-Smaug-LLama-70b

    parameters:

      weight: 0.28

      density: 0.90

      gamma: 0.01

  - model: \Llama-3-Lumimaid-70B-v0.1-alt

    parameters:

      weight: 0.15

      density: 0.90

      gamma: 0.01

  - model: \Tess-2.0-Llama-3-70B-v0.2

    parameters:

      weight: 0.06

      density: 0.90

      gamma: 0.01

  - model: \Smaug-Llama-3-70B-Instruct

    parameters:

      weight: 0.04

      density: 0.90

      gamma: 0.01

  - model: \Llama-3-70B-Synthia-v3.5

    parameters:

      weight: 0.05

      density: 0.90

      gamma: 0.01

  - model: \Llama-3-70B-Instruct-Gradient-524k

    parameters:

      weight: 0.03

      density: 0.90

      gamma: 0.01

  - model: \Meta-LLama-3-Cat-A-LLama-70b

    parameters:

      weight: 0.14

      density: 0.90

      gamma: 0.01

merge_method: breadcrumbs_ties

base_model: I:\Llama-3-70B-Instruct-Gradient-262k
dtype: float16
```