Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,9 @@ tags:
|
|
10 |
|
11 |
# WizardLM-2-4x7B-MoE
|
12 |
|
13 |
-
WizardLM-2-4x7B-MoE is an experimental MoE model made with [Mergekit](https://github.com/arcee-ai/mergekit). It was made by combining four [WizardLM-2-7B](https://huggingface.co/microsoft/WizardLM-2-7B) models using the random gate mode.
|
|
|
|
|
14 |
|
15 |
# Quanitized versions
|
16 |
|
|
|
10 |
|
11 |
# WizardLM-2-4x7B-MoE
|
12 |
|
13 |
+
WizardLM-2-4x7B-MoE is an experimental MoE model made with [Mergekit](https://github.com/arcee-ai/mergekit). It was made by combining four [WizardLM-2-7B](https://huggingface.co/microsoft/WizardLM-2-7B) models using the random gate mode.
|
14 |
+
|
15 |
+
Please be sure to set experts per token to 4 for the best results! Context length should be the same as Mistral-7B-Instruct-v0.1 (8k tokens). For instruction templates, Vicuna-v1.1 is recommended.
|
16 |
|
17 |
# Quanitized versions
|
18 |
|