Edit model card
Qwen2 fine-tune

MaziyarPanahi/Qwen2-72B-Instruct-v0.1

This is a fine-tuned version of the Qwen/Qwen2-72B-Instruct model. It aims to improve the base model across all benchmarks.

⚑ Quantized GGUF

All GGUF models are available here: MaziyarPanahi/Qwen2-72B-Instruct-v0.1-GGUF

πŸ† Open LLM Leaderboard Evaluation Results

coming soon!

Tasks Version Filter n-shot Metric Value Stderr
truthfulqa_mc2 2 none 0 acc 0.6761 Β± 0.0148
Tasks Version Filter n-shot Metric Value Stderr
winogrande 1 none 5 acc 0.8248 Β± 0.0107
Tasks Version Filter n-shot Metric Value Stderr
arc_challenge 1 none 25 acc 0.6852 Β± 0.0136
none 25 acc_norm 0.7184 Β± 0.0131
Tasks Version Filter n-shot Metric Value Stderr
gsm8k 3 strict-match 5 exact_match 0.8582 Β± 0.0096
flexible-extract 5 exact_match 0.8893 Β± 0.0086

Prompt Template

This model uses ChatML prompt template:

<|im_start|>system
{System}
<|im_end|>
<|im_start|>user
{User}
<|im_end|>
<|im_start|>assistant
{Assistant}

How to use


# Use a pipeline as a high-level helper

from transformers import pipeline

messages = [
    {"role": "user", "content": "Who are you?"},
]
pipe = pipeline("text-generation", model="MaziyarPanahi/Qwen2-72B-Instruct-v0.1")
pipe(messages)


# Load model directly

from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("MaziyarPanahi/Qwen2-72B-Instruct-v0.1")
model = AutoModelForCausalLM.from_pretrained("MaziyarPanahi/Qwen2-72B-Instruct-v0.1")
Downloads last month
8
Safetensors
Model size
72.7B params
Tensor type
BF16
Β·
Inference API
Input a message to start chatting with MaziyarPanahi/Qwen2-72B-Instruct-v0.1.
Inference API (serverless) has been turned off for this model.

Finetuned from

Collections including MaziyarPanahi/Qwen2-72B-Instruct-v0.1