Downtown-Case/jukofyork_creative-writer-32b-preview-exl2-4.0bpw Text Generation • Updated 11 days ago • 8
24GB VRAM Optimal Quants Collection When asked what I use locally on a 24GB card, this is what I point to. I favor exl2s for long context, GGUF for very short context. • 12 items • Updated Oct 31, 2024 • 3
Downtown-Case/EVA-UNIT-01_EVA-Qwen2.5-32B-v0.1-exl2-4.1bpw Text Generation • Updated Oct 31, 2024 • 11