Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

mistral-7b-dolphin-sft - GGUF

Original model description:

license: apache-2.0

mistral-7b-dolphin-sft

mistral-7b-dolphin-sft is an SFT fine-tuned version of unsloth/mistral-7b-bnb-4bit using the cognitivecomputations/dolphin dataset.

Fine-tuning configuration

LoRA

  • r: 256
  • LoRA alpha: 128
  • LoRA dropout: 0.0

Training arguments

  • Epochs: 1
  • Batch size: 4
  • Gradient accumulation steps: 6
  • Optimizer: adamw_torch_fused
  • Max steps: 100
  • Learning rate: 0.0002
  • Weight decay: 0.1
  • Learning rate scheduler type: linear
  • Max seq length: 2048
  • 4-bit bnb: True

Trained with Unsloth and Huggingface's TRL library.

Downloads last month
258
GGUF
Model size
7.24B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .