This is a more recent quantized GGUF version of OpenOrca-Platypus-13B quantized to 4_K-M respecting the new format from (https://github.com/ggerganov/llama.cpp).

(link to the original model : https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B)

Instruction template :

### Instruction:

<prompt> (without the <>)

### Response:
Downloads last month
5
GGUF
Model size
13B params
Architecture
llama

4-bit

Inference API
Unable to determine this model's library. Check the docs .

Dataset used to train kroonen/OpenOrca-Platypus2-13B-GGUF