Edit model card

MixTAO-7Bx2-MoE

MixTAO-7Bx2-MoE is a Mixture of Experts (MoE). This model is mainly used for large model technology experiments, and increasingly perfect iterations will eventually create high-level large language models.

Prompt Template (Alpaca)

### Instruction:
<prompt> (without the <>)
### Response:

πŸ¦’ Colab

Link Info - Model Name
Open In Colab MixTAO-7Bx2-MoE-v8.1
mixtao-7bx2-moe-v8.1.Q4_K_M.gguf GGUF of MixTAO-7Bx2-MoE-v8.1
Only Q4_K_M in https://huggingface.co/zhengr/MixTAO-7Bx2-MoE-v8.1-GGUF
Demo Space https://huggingface.co/spaces/zhengr/MixTAO-7Bx2-MoE-v8.1/

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 77.50
AI2 Reasoning Challenge (25-Shot) 73.81
HellaSwag (10-Shot) 89.22
MMLU (5-Shot) 64.92
TruthfulQA (0-shot) 78.57
Winogrande (5-shot) 87.37
GSM8k (5-shot) 71.11
Downloads last month
7,245
Safetensors
Model size
12.9B params
Tensor type
BF16
Β·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Space using zhengr/MixTAO-7Bx2-MoE-v8.1 1

Evaluation results