Edit model card

gemma-2-9b GGUF

Llama.cpp version b3259 was used for hf to gguf conversion.

Original model: https://huggingface.co/google/gemma-2-9b

Available Precisions:

  • f16
  • q8_0

License

Gemma Terms of Use applies the same as the original model.

Downloads last month
28
GGUF
Model size
9.24B params
Architecture
gemma2

8-bit

16-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.