I am trying to convert the model into BitsAndBytes with NF4 (4-bit)
#1
by
sabaridsnfuji
- opened
Could you guide me on how to save a fine-tuned LoRA model for Llama-3.2-11B-Vision-Instruct in 4-bit precision for optimized inference, similar to the repository SeanScripts/Llama-3.2-11B-Vision-Instruct-nf4 ?