This is an experimental 2x8B moe with random gates, using the following 2 models

Important

Make sure to add </s> a stop sequence as it uses llama-3-cat-8B-instruct-V1 as the base model.

Update:

Due to request i decided to add the rest of the quants. Enjoy

Mergekit recipe of the model if too lazy to check the files:

base_model: TheSkullery/llama-3-cat-8b-instruct-v1
gate_mode: random
dtype: bfloat16
experts_per_token: 2
experts:
 - source_model: TheSkullery/llama-3-cat-8b-instruct-v1
   positive_prompts:
    - " "
 - source_model: NousResearch/Hermes-2-Theta-Llama-3-8B
   positive_prompts:
    - " "
Downloads last month
41
GGUF
Model size
13.7B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

16-bit

Inference API
Unable to determine this model's library. Check the docs .