โš›๏ธ Q Model: Optimized for Enhanced Quantized Inference Capability

This model has been specially optimized to improve the performance of quantized inference and is recommended for use in 3 to 8-bit quantization scenarios.

Downloads last month
298
GGUF
Model size
70.6B params
Architecture
llama

2-bit

3-bit

4-bit

Inference API
Unable to determine this model's library. Check the docs .

Model tree for OpenBuddy/openbuddy-llama3.3-70b-v24.2q-gguf

Quantized
(3)
this model