Main page: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981


Update:

Always use V2 by default.

V2 is quantized in a better way to turn off the second stage of double quant.

V2 is 0.5 GB larger than the previous version, since the chunk 64 norm is now stored in full precision float32, making it much more precise than the previous version. Also, since V2 does not have second compression stage, it now has less computation overhead for on-the-fly decompression, making the inference a bit faster.

The only drawback of V2 is being 0.5 GB larger.


Main model in bnb-nf4 (v1 with chunk 64 norm in nf4, v2 with chunk 64 norm in float32)

T5xxl in fp8e4m3fn

CLIP-L in fp16

VAE in bf16

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model's library. Check the docs .

Model tree for lllyasviel/flux1-dev-bnb-nf4

Finetunes
1 model

Spaces using lllyasviel/flux1-dev-bnb-nf4 15