stable-diffusion-2-1-GGUF

Original Model

stabilityai/stable-diffusion-2-1

Run with sd-api-server

Go to the sd-api-server repository for more information.

Quantized GGUF Models

Using formats of different precisions will yield results of varying quality.

f32 f16 q8_0 q5_0 q5_1 q4_0 q4_1
Name Quant method Bits Size Use case
v2-1_768-nonema-pruned-Q4_0.gguf Q4_0 2 1.70 GB
v2-1_768-nonema-pruned-Q4_1.gguf Q4_1 3 1.74 GB
v2-1_768-nonema-pruned-Q5_0.gguf Q5_0 3 1.78 GB
v2-1_768-nonema-pruned-Q5_1.gguf Q5_1 3 1.82 GB
v2-1_768-nonema-pruned-Q8_0.gguf Q8_0 4 2.01 GB
v2-1_768-nonema-pruned-f16.gguf f16 4 2.61 GB
v2-1_768-nonema-pruned-f32.gguf f32 4 5.21 GB
Downloads last month
540
GGUF
Model size
1.3B params
Architecture
undefined

4-bit

5-bit

8-bit

16-bit

32-bit

Inference Examples
Unable to determine this model's library. Check the docs .

Model tree for second-state/stable-diffusion-2-1-GGUF

Quantized
(1)
this model