- Original model is spow12/Ko-Qwen2-7B-Instruct
- quantized using llama.cpp - b3510
<|im_start|>system
{System}<|im_end|>
<|im_start|>user
{User}<|im_end|>
<|im_start|>assistant
{Assistant}
"Flash Attention" function must be activated. why?
- Downloads last month
- 124
Model tree for joongi007/Ko-Qwen2-7B-Instruct-GGUF
Base model
spow12/Ko-Qwen2-7B-Instruct