Air-Striker-Mixtral-8x7B-ZLoss

Experimental model, trained using config and Transformers/Axolotl forks provided by Doctor-Shotgun

Model was fine-tuned from Mixtral-8x7B-v0.1 with airoboros-3.2 dataset, for 4 epochs, ChatML prompt format at 8K context length.

Downloads last month
39
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.

Model tree for LoneStriker/Air-Striker-Mixtral-8x7B-ZLoss-LoRA

Merges
2 models

Dataset used to train LoneStriker/Air-Striker-Mixtral-8x7B-ZLoss-LoRA