miiqu-105b-v1.0
Developed by Infinimol AI GmbH
Also Available:
8th place on EQ-Bench, beating Qwen1.5-72B-Chat, miqudev/miqu-1-70b, mistral-medium and claude-3-sonnet-20240229. All without fine-tuning or additional training.
Thanks for support from: turboderp, silphendio, sqrkl, and ngxson!
Model Details
- Max Context: 32768 tokens
- Layers: 105
Prompt template: ChatML or Mistral
chatml:
<|im_start|><|user|>\n<|user-message|><|im_end|>\n<|im_start|><|bot|>\n<|bot-message|><|im_end|>\n
mistral:
[INST] <|user|><|user-message|>[/INST]<|bot|><|bot-message|></s>
- Downloads last month
- 0
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.