Fp16 -> GGUF 8B script?
1
#40 opened about 1 month ago
by
vdruts
Not able to download it
1
#38 opened about 2 months ago
by
iffishells
Create an equivalent to GGUF for Diffusers models?
1
#37 opened about 2 months ago
by
julien-c
FP16 (safetensors) & Q8_0 (gguf) example PNG (anime girl) comparison (with ComfyUI workflow!)
7
#36 opened 2 months ago
by
fireYtail
What's the python code for utilizing this quantized model?
#35 opened 3 months ago
by
enexorb
Create config.json
#34 opened 3 months ago
by
Jash-RA
video
#33 opened 3 months ago
by
goodasdgood
run
1
#32 opened 3 months ago
by
goodasdgood
Is it using ggml to compute?
1
#30 opened 4 months ago
by
CHNtentes
FLUX GGUF conversion
1
#29 opened 4 months ago
by
bsingh1324
cant see the Q5_K_M gguf quant
2
#28 opened 4 months ago
by
Ai11Ali
Q6_K does not play as well with LoRAs
1
#25 opened 4 months ago
by
SubtleOne
Q8 vs Q4_1 vs FP8 on a 1070ti
#23 opened 4 months ago
by
gfl1973
Which one is the best?
12
#21 opened 4 months ago
by
akiokawahito
Can you add Q4K_m please ?
1
#19 opened 4 months ago
by
mirek190
code for use this quantized model
3
#18 opened 4 months ago
by
Mahdimohseni0333
Compared to the regular FP8 model, what is the better performance of the 8BIT model here
4
#16 opened 4 months ago
by
demo001s
all K quants comparison using fp16/fp8 t5
16
#15 opened 4 months ago
by
Nelathan
What is the _K on the new files ?
1
#14 opened 4 months ago
by
Fabienne
Can it use LoRAs?
2
#13 opened 4 months ago
by
zdeto
Can you add a link to that "install instructions GitHub" ??? And possible a super simple ComfyUI demo workflow?
1
#12 opened 4 months ago
by
AlexData257
Please explain the difference between the two models
3
#11 opened 4 months ago
by
martjay
Can this kind of merged model be converted to GGUF?
2
#10 opened 4 months ago
by
kakaxi
My experience
9
#9 opened 4 months ago
by
SaisExperiments
k-quants possible?
5
#2 opened 4 months ago
by
CHNtentes