Edit model card
Configuration Parsing Warning: In config.json: "quantization_config.bits" must be an integer
Configuration Parsing Warning: In config.json: "quantization_config.bits" must be less than or equal to 8

8.1 bpw exl2 quant of Smaug-Llama-3-70B-Instruct

Downloads last month
2
Inference API
Input a message to start chatting with llmixer/Smaug-Llama-3-70B-Instruct-8.1bpw-h8-exl2.
Model is too large to load in Inference API (serverless). To try the model, launch it on Inference Endpoints (dedicated) instead.