Model Card for Model ID
Model Details
Model Description
GPTQ-4bit quantization version (use exllamav2) of https://huggingface.co/T3LS/stella-mrl-large-zh-v3.5-1792d-1024
Uses
model = AutoModelForSequenceClassification.from_pretrained(
'T3LS/stella-mrl-large-zh-v3.5-1792d-1024-gptq-4bit',
device_map='cuda' # Exllamav2 backend requires all the modules to be on GPU
)
- Downloads last month
- 10
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.