This repository hosts quantized versions of the Mistral-Nemo-Instruct-2407 model.
Format: GGUFConverter: llama.cpp 3a14e00366399040a139c67dd5951177a8cb5695Quantizer: LM-Kit.NET 2024.7.11
For more detailed information on the base model, please visit the following link:
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit