Edit model card

Model Card for Model ID

MistralAI 7B model fine-tuned for 1 epoch on Dataricks instruction tuning dataset.

Model Details

Model Description

Mistral Repository

Training Details

Training Data

Training Procedure

  • Quantized Low-Rank Adaptation (QLoRA)
  • Transformers Trainer
  • DataCollatorForSeq2Seq
  • Distributed Data Parallel (DDP) across two GPUs

Preprocessing

Manually created tokenized 'labels' for the dataset. Prompt template utilized basic template for instruction-tuning

Hardware

Performed fine-tuning with 2 * A100 GPUs

  • Provided by Gnewsoft during work period Model and dataset are too large for free run sessions on Google Colab
Downloads last month
14
Safetensors
Model size
7.24B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train Chahnwoo/Mistral-7B-v0.1-1E-QLoRA-SFT-Test