A quantized version of Granite Guardian 3.1 2B model from IBM.

Quantization is done by llama.cpp.

Model Summary (from original repository)

Granite Guardian 3.1 2B is a fine-tuned Granite 3.1 2B Instruct model designed to detect risks in prompts and responses. It can help with risk detection along many key dimensions catalogued in the IBM AI Risk Atlas. It is trained on unique data comprising human annotations and synthetic data informed by internal red-teaming. It outperforms other open-source models in the same space on standard benchmarks.

Downloads last month
17
GGUF
Model size
2.53B params
Architecture
granite

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for ktoprakucar/granite-guardian-3.1-2b-Q8-GGUF

Quantized
(3)
this model