metadata
tags:
- quantized
- 4-bit
- AWQ
- autotrain_compatible
- endpoints_compatible
- text-generation-inference
license: apache-2.0
language:
- en
base_model: Vezora/Mistral-22B-v0.1
model_creator: Vezora
model_name: Mistral-22B-v0.1
model_type: mistral
pipeline_tag: text-generation
inference: false
Vezora/Mistral-22B-v0.1 AWQ
- Model creator: Vezora
- Original model: Mistral-22B-v0.1
Model Summary
This model is not an moe, it is infact a 22B parameter dense model!
Just one day after the release of Mixtral-8x-22b, we are excited to introduce our handcrafted experimental model, Mistral-22b-V.01. This model is a culmination of equal knowledge distilled from all experts into a single, dense 22b model. This model is not a single trained expert, rather its a compressed MOE model, turning it into a dense 22b mode. This is the first working MOE to Dense model conversion.
How to use
GUANACO PROMPT FORMAT YOU MUST USE THE GUANACO PROMPT FORMAT SHOWN BELOW. Not using this prompt format will lead to sub optimal results.
- This model requires a specific chat template, as the training format was Guanaco this is what it looks like:
- "### System: You are a helpful assistant. ### Human###: Give me the best chili recipe you can ###Assistant: Here is the best chili recipe..."