Text Generation
Safetensors

GPTQ Quantization

#1
by CyberTimon - opened

Hello there!

Thank you for this amazing model. Can you maybe provide a quantized model with GPTQ?
Thank you very much.

Kind regards,
Timon Käch

Glad you like it. I tried gptq already on a model merged with the adapter, but it said something like llama was expecting an llm.header layer and then instantiated the layer itself. The same for awq.
The model that came out of that only generated rubbish, currently I don't know how to circumvent this issue.

I asked @TheBloke to help out

KnutJaegersberg changed discussion status to closed

Sign up or log in to comment