4bit GPTQ version?
Thanks for training this dataset. I'd love to try it but was curious if you could upload a 4-bit quantized version of this model?
(It would take a lot less space and VRAM). Thanks!
Thanks for training this dataset. I'd love to try it but was curious if you could upload a 4-bit quantized version of this model?
(It would take a lot less space and VRAM). Thanks!
Yes I'll do so right now
Would it be possible to also get a version that works with llama.cpp (https://github.com/ggerganov/llama.cpp)?
Would it be possible to also get a version that works with llama.cpp (https://github.com/ggerganov/llama.cpp)?
I'm sorry I've never worked with it, so I'm not quite sure how to
Thanks for training this dataset. I'd love to try it but was curious if you could upload a 4-bit quantized version of this model?
(It would take a lot less space and VRAM). Thanks!
I'm uploading a version 2 of the RP Lora merged with GPT4-x-Alpaca model in 4bit quantization right now, you can get it here:
https://huggingface.co/teknium/GPT4-x-Alpaca13b-RolePlayLora-4bit-v2
Would it be possible to also get a version that works with llama.cpp (https://github.com/ggerganov/llama.cpp)?
I'm sorry I've never worked with it, so I'm not quite sure how to
I can take a crack at it. It's for compatibility with Llama.cpp