Problem in running with vllm

#4
by babakgh - opened

Hi
Command vllm serve "bartowski/Llama-3.3-70B-Instruct-GGUF" returns following error. how to fix this?

raise ValueError(f"No supported config format found in {model}")
ValueError: No supported config format found in bartowski/Llama-3.3-70B-Instruct-GGUF

i think you need to add a config.json file with:

{
    "model_type": "llama"
}

can you confirm if this works?

Hi,

I'm having the same issue. I tried downloading the Llama-3.3-70B-Instruct-Q6_K_L file as well, and I combined them using llama-gguf-split --merge. The file is fine as I tried loading it with llama-cpp-python and it works. When I run

vllm serve Llama3-3.3-70B-Instruct-Q6_K_L.gguf --tokenizer=meta-llama/Llama-3.3-70B-Instruct --api-key=<KEY> --max-model-len=32000

I get

ValueError: No supported config format found in Llama3-3.3-70B-Instruct-Q6_K_L.gguf

I tried putting a config.json file in the same directory as the model with the elements you said above but it doesn't work. Please advise. Thanks!

Sign up or log in to comment