|
--- |
|
base_model: black-forest-labs/FLUX.1-dev |
|
library_name: gguf |
|
license: other |
|
license_name: flux-1-dev-non-commercial-license |
|
license_link: LICENSE.md |
|
quantized_by: mo137 |
|
tags: |
|
- text-to-image |
|
- image-generation |
|
- flux |
|
--- |
|
|
|
# ⚠️ bf16 GGUF Flux is not yet supported by any UI like ComfyUI or Forge! ⚠️ |
|
|
|
Flux.1-dev in **bfloat16**, exactly the same tensors as in black-forest-labs' original safetensors. |
|
AFAIK no inference software supports bf16 FLUX yet, but maybe this file can help. |
|
I have no clue if it would improve anything, but bf16 –> fp16 conversion is lossy, so... maybe? |
|
|
|
This is a direct GGUF conversion of [black-forest-labs/FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev/tree/main) |
|
|
|
As this is a quantized model not a finetune, all the same restrictions/original license terms still apply. |
|
|
|
The model files can be used with the [ComfyUI-GGUF](https://github.com/city96/ComfyUI-GGUF) custom node. |
|
|
|
Place model files in `ComfyUI/models/unet` - see the GitHub readme for further install instructions. |
|
|
|
Please refer to [this chart](https://github.com/ggerganov/llama.cpp/blob/master/examples/perplexity/README.md#llama-3-8b-scoreboard) for a basic overview of quantization types. |
|
|
|
(Model card mostly copied from [city96/FLUX.1-dev-gguf](https://huggingface.co/city96/FLUX.1-dev-gguf) - which contains conventional and useful GGUF files.) |
|
|