Edit model card

Overview

Fine-tuned Llama-2 7B with an uncensored/unfiltered Wizard-Vicuna conversation dataset (originally from ehartford/wizard_vicuna_70k_unfiltered). Used QLoRA for fine-tuning. Trained for one epoch on a 24GB GPU (NVIDIA A10G) instance, took ~19 hours to train.

The version here is the fp16 HuggingFace model.

GGML & GPTQ versions

Thanks to TheBloke, he has created the GGML and GPTQ versions:

Running in Ollama

https://ollama.com/library/llama2-uncensored

Prompt style

The model was trained with the following prompt style:

### HUMAN:
Hello

### RESPONSE:
Hi, how are you?

### HUMAN:
I'm fine.

### RESPONSE:
How can I help you?
...

Training code

Code used to train the model is available here.

To reproduce the results:

git clone https://github.com/georgesung/llm_qlora
cd llm_qlora
pip install -r requirements.txt
python train.py configs/llama2_7b_chat_uncensored.yaml

Fine-tuning guide

https://georgesung.github.io/ai/qlora-ift/

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 43.39
ARC (25-shot) 53.58
HellaSwag (10-shot) 78.66
MMLU (5-shot) 44.49
TruthfulQA (0-shot) 41.34
Winogrande (5-shot) 74.11
GSM8K (5-shot) 5.84
DROP (3-shot) 5.69
Downloads last month
9,393
Safetensors
Model size
6.74B params
Tensor type
F32
Β·
Inference API
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.

Dataset used to train georgesung/llama2_7b_chat_uncensored

Spaces using georgesung/llama2_7b_chat_uncensored 57