metadata
language:
- pl
license: apache-2.0
library_name: transformers
tags:
- finetuned
- gguf
- 4bit
inference: false
pipeline_tag: text-generation
base_model: speakleash/Bielik-11B-v2.2-Instruct
Bielik-11B-v2.2-Instruct-Quanto-4bit
This model was converted to Quanto format from SpeakLeash's Bielik-11B-v.2.2-Instruct.
DISCLAIMER: Be aware that quantised models show reduced response quality and possible hallucinations!
About Quanto
Optimum Quanto is a pytorch quantization backend for optimum. Model can be loaded using:
from optimum.quanto import QuantizedModelForCausalLM
qmodel = QuantizedModelForCausalLM.from_pretrained('speakleash/Bielik-11B-v2.2-Instruct-Quanto-4bit')
Model description:
- Developed by: SpeakLeash & ACK Cyfronet AGH
- Language: Polish
- Model type: causal decoder-only
- Quant from: Bielik-11B-v2.2-Instruct
- Finetuned from: Bielik-11B-v2
- License: Apache 2.0 and Terms of Use
Responsible for model quantization
- Remigiusz KinasSpeakLeash - team leadership, conceptualizing, calibration data preparation, process creation and quantized model delivery.
Contact Us
If you have any questions or suggestions, please use the discussion tab. If you want to contact us directly, join our Discord SpeakLeash.