Model Info

Layers Context Template
40
16384
I think it's Alpaca. ChatML seems to work too but answers are probably worse?

Below is Alpaca template. I think there should be a new line (\n) after ### Response:. Sysprompt which is the first line can be changed of course:

You are an exceptionally intelligent coding assistant that consistently delivers accurate and reliable responses to user instructions.

### Instruction:
{instruction}

### Response:

Quantization info

Without imatrix. Quantized with llama.cpp b2333 (2024-03-04).

Downloads last month
53
GGUF
Model size
16B params
Architecture
starcoder2

3-bit

4-bit

5-bit

6-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model authors have turned it off explicitly.

Model tree for arzeth/opencsg-starcoder2-15b-v0.1-GGUF

Quantized
(4)
this model