divinetaco commited on
Commit
ba80e0d
1 Parent(s): d953764

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +35 -3
  2. measurement.json +0 -0
  3. mergekit_config.yml +10 -0
README.md CHANGED
@@ -1,3 +1,35 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - alpindale/WizardLM-2-8x22B
4
+ - openbmb/Eurux-8x22b-nca
5
+ - openbmb/Eurux-8x22b-kto
6
+ - fireworks-ai/mixtral-8x22b-instruct-oh
7
+ - migtissera/Tess-2.0-Mixtral-8x22B
8
+ - mistralai/Mixtral-8x22B-v0.1
9
+ tags:
10
+ - mergekit
11
+ - merge
12
+ ---
13
+ # WizardLM-2-8x22B-model_stock
14
+
15
+ A [mergekit](https://github.com/arcee-ai/mergekit) model_stock merge made with the aim of improving WizardLM-2-8x22B.
16
+ The resulting model suppresses WizardLM-2-8x22B's overly flowery and positive writing style whilst retaining useful features such as CoT. Extremely coherant even at long contexts, and benched above the WLM-2 base in intelligence tests.
17
+ Use vicuna prompt as per WizardLM-2-8x22B base.
18
+
19
+ exllamav2 [measurement.json](./measurement.json)
20
+
21
+ mergekit_config.yml
22
+ ```yml
23
+ models:
24
+ - model: alpindale/WizardLM-2-8x22B
25
+ - model: openbmb/Eurux-8x22b-kto
26
+ - model: openbmb/Eurux-8x22b-nca
27
+ - model: mistralai/Mixtral-8x22B-v0.1
28
+ - model: migtissera/Tess-2.0-Mixtral-8x22B
29
+ - model: fireworks-ai/mixtral-8x22b-instruct-oh
30
+ base_model: alpindale/WizardLM-2-8x22B
31
+ merge_method: model_stock
32
+ dtype: bfloat16
33
+ ```
34
+
35
+ Likely won't be uploading the full weights myself due to bandwidth limitations.
measurement.json ADDED
The diff for this file is too large to render. See raw diff
 
mergekit_config.yml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - model: WizardLM-2-8x22B
3
+ - model: Eurux-8x22b-kto
4
+ - model: Mixtral-8x22B-v0.1
5
+ - model: Tess-2.0-Mixtral-8x22B
6
+ - model: mixtral-8x22b-instruct-oh
7
+ - model: Eurux-8x22b-nca
8
+ base_model: WizardLM-2-8x22B
9
+ merge_method: model_stock
10
+ dtype: bfloat16