Text Generation
Transformers
Safetensors
English
llama
llama-factory
Not-For-All-Audiences
conversational
text-generation-inference
Inference Endpoints

My quantizations.

#1
by ZeroWw - opened

These are my own quantizations (updated almost daily).

The difference with normal quantizations is that I quantize the output and embed tensors to f16.
and the other tensors to 15_k,q6_k or q8_0.
This creates models that are little or not degraded at all and have a smaller size.
They run at about 3-6 t/sec on CPU only using llama.cpp
And obviously faster on computers with potent GPUs

Nothing is Real org

Thanks! I've added the link to the model card

AuriAetherwiing changed discussion status to closed
aaronday3 changed discussion status to open
aaronday3 changed discussion status to closed

Thanks! I've added the link to the model card

the first link was wrong I corrected it.

Sign up or log in to comment