Update README.md
Browse files
README.md
CHANGED
@@ -6,7 +6,7 @@ tags:
|
|
6 |
- merge
|
7 |
license: apache-2.0
|
8 |
---
|
9 |
-
![thumbnail](https://huggingface.co/mayacinka/NeuralZephyr-Beagle-7B/blob/main/
|
10 |
# merge
|
11 |
|
12 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
@@ -14,13 +14,13 @@ This is a merge of pre-trained language models created using [mergekit](https://
|
|
14 |
## Merge Details
|
15 |
### Merge Method
|
16 |
|
17 |
-
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using NeuralTrix-7B-dpo as a base.
|
18 |
|
19 |
### Models Merged
|
20 |
|
21 |
The following models were included in the merge:
|
22 |
-
* NeuralBeagle14-7B
|
23 |
-
* zephyr-7b-alpha
|
24 |
|
25 |
### Configuration
|
26 |
|
@@ -28,17 +28,17 @@ The following YAML configuration was used to produce this model:
|
|
28 |
|
29 |
```yaml
|
30 |
models:
|
31 |
-
- model: NeuralTrix-7B-dpo
|
32 |
-
- model: zephyr-7b-alpha
|
33 |
parameters:
|
34 |
density: 0.83
|
35 |
weight: 0.4
|
36 |
-
- model: NeuralBeagle14-7B
|
37 |
parameters:
|
38 |
density: 0.83
|
39 |
weight: 0.6
|
40 |
merge_method: dare_ties
|
41 |
-
base_model: NeuralTrix-7B-dpo
|
42 |
parameters:
|
43 |
int8_mask: true
|
44 |
dtype: bfloat16
|
|
|
6 |
- merge
|
7 |
license: apache-2.0
|
8 |
---
|
9 |
+
![thumbnail](https://huggingface.co/mayacinka/NeuralZephyr-Beagle-7B/blob/main/thumbn.png)
|
10 |
# merge
|
11 |
|
12 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
|
|
14 |
## Merge Details
|
15 |
### Merge Method
|
16 |
|
17 |
+
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using CultriX/NeuralTrix-7B-dpo as a base.
|
18 |
|
19 |
### Models Merged
|
20 |
|
21 |
The following models were included in the merge:
|
22 |
+
* mlabonne/NeuralBeagle14-7B
|
23 |
+
* HuggingFaceH4/zephyr-7b-alpha
|
24 |
|
25 |
### Configuration
|
26 |
|
|
|
28 |
|
29 |
```yaml
|
30 |
models:
|
31 |
+
- model: CultriX/NeuralTrix-7B-dpo
|
32 |
+
- model: HuggingFaceH4/zephyr-7b-alpha
|
33 |
parameters:
|
34 |
density: 0.83
|
35 |
weight: 0.4
|
36 |
+
- model: mlabonne/NeuralBeagle14-7B
|
37 |
parameters:
|
38 |
density: 0.83
|
39 |
weight: 0.6
|
40 |
merge_method: dare_ties
|
41 |
+
base_model: CultriX/NeuralTrix-7B-dpo
|
42 |
parameters:
|
43 |
int8_mask: true
|
44 |
dtype: bfloat16
|