NeuralNovel commited on
Commit
4a9eb91
1 Parent(s): 45dd19b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -10,16 +10,19 @@ tags:
10
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/645cfe4603fc86c46b3e46d1/1Uoxp_Bl9UwF-1K4KzwVN.png)
11
 
12
  # ConvexAI/Solutus-3x7B
13
- [Join our Discord!](https://discord.gg/XFChPkGd)
14
 
15
  A model to test how MoE will route without square expansion.
16
 
 
 
17
  The config looks like this...(detailed version is in the files and versions):
18
  - [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
19
  - [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
20
  - [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
21
  - [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
22
 
 
 
23
  # "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
24
  ### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
25
 
 
10
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/645cfe4603fc86c46b3e46d1/1Uoxp_Bl9UwF-1K4KzwVN.png)
11
 
12
  # ConvexAI/Solutus-3x7B
 
13
 
14
  A model to test how MoE will route without square expansion.
15
 
16
+ [Join our Discord!](https://discord.gg/rJXGjmxqzS)
17
+
18
  The config looks like this...(detailed version is in the files and versions):
19
  - [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
20
  - [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
21
  - [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
22
  - [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
23
 
24
+
25
+
26
  # "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
27
  ### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
28