โ๏ธ Q Model: Optimized for Enhanced Quantized Inference Capability
This model has been specially optimized to improve the performance of quantized inference and is recommended for use in 3 to 8-bit quantization scenarios.
Downloads last month
298
GGUF
Model size
70.6B params
Architecture
llama
2-bit
3-bit
4-bit
Inference API
Unable to determine this model's library. Check the
docs
.
Model tree for OpenBuddy/openbuddy-llama3.3-70b-v24.2q-gguf