Cannot use the model

#2
by Daaku-C5 - opened

Hi, I'm new to quantization. I'm trying to use the Q6_K GGUF file with the below code but I'm getting this error. Can you please help in resolving this issue. I'm running this in 8 GB of GPU

from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "bartowski/granite-3.0-8b-instruct-GGUF"
filename = "granite-3.0-8b-instruct-Q6_K_L.gguf"
tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename)
model = AutoModelForCausalLM.from_pretrained(model_id, gguf_file=filename)

ERROR:

ValueError: Architecture granite not supported

I'm following this documentation from HF https://huggingface.co/docs/transformers/main/en/gguf

Sign up or log in to comment