mayacinka commited on
Commit
cadcbc8
·
verified ·
1 Parent(s): 8934290

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -8
README.md CHANGED
@@ -6,7 +6,7 @@ tags:
6
  - merge
7
  license: apache-2.0
8
  ---
9
- ![thumbnail](https://huggingface.co/mayacinka/NeuralZephyr-Beagle-7B/blob/main/thumbnail.webp)
10
  # merge
11
 
12
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
@@ -14,13 +14,13 @@ This is a merge of pre-trained language models created using [mergekit](https://
14
  ## Merge Details
15
  ### Merge Method
16
 
17
- This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using NeuralTrix-7B-dpo as a base.
18
 
19
  ### Models Merged
20
 
21
  The following models were included in the merge:
22
- * NeuralBeagle14-7B
23
- * zephyr-7b-alpha
24
 
25
  ### Configuration
26
 
@@ -28,17 +28,17 @@ The following YAML configuration was used to produce this model:
28
 
29
  ```yaml
30
  models:
31
- - model: NeuralTrix-7B-dpo
32
- - model: zephyr-7b-alpha
33
  parameters:
34
  density: 0.83
35
  weight: 0.4
36
- - model: NeuralBeagle14-7B
37
  parameters:
38
  density: 0.83
39
  weight: 0.6
40
  merge_method: dare_ties
41
- base_model: NeuralTrix-7B-dpo
42
  parameters:
43
  int8_mask: true
44
  dtype: bfloat16
 
6
  - merge
7
  license: apache-2.0
8
  ---
9
+ ![thumbnail](https://huggingface.co/mayacinka/NeuralZephyr-Beagle-7B/blob/main/thumbn.png)
10
  # merge
11
 
12
  This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
 
14
  ## Merge Details
15
  ### Merge Method
16
 
17
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using CultriX/NeuralTrix-7B-dpo as a base.
18
 
19
  ### Models Merged
20
 
21
  The following models were included in the merge:
22
+ * mlabonne/NeuralBeagle14-7B
23
+ * HuggingFaceH4/zephyr-7b-alpha
24
 
25
  ### Configuration
26
 
 
28
 
29
  ```yaml
30
  models:
31
+ - model: CultriX/NeuralTrix-7B-dpo
32
+ - model: HuggingFaceH4/zephyr-7b-alpha
33
  parameters:
34
  density: 0.83
35
  weight: 0.4
36
+ - model: mlabonne/NeuralBeagle14-7B
37
  parameters:
38
  density: 0.83
39
  weight: 0.6
40
  merge_method: dare_ties
41
+ base_model: CultriX/NeuralTrix-7B-dpo
42
  parameters:
43
  int8_mask: true
44
  dtype: bfloat16