Optimised Quants for high-throughput deployments! Compatible with Transformers, TGI & VLLM 🤗
-
hugging-quants/Meta-Llama-3.1-405B-Instruct-AWQ-INT4
Text Generation • Updated • 10.2k • 32 -
hugging-quants/Meta-Llama-3.1-405B-Instruct-BNB-NF4
Text Generation • Updated • 248 • 4 -
hugging-quants/Meta-Llama-3.1-405B-Instruct-GPTQ-INT4
Text Generation • Updated • 916 • 15 -
hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4
Text Generation • Updated • 68k • 69