Description
Converted to f16 using llama_cpp convert.py script, then quantized to q6_K using quantize from the same llama_cpp repository.
Resulting file was split into 2 parts.
Note: HF does not support uploading files larger than 50GB.
File require joining
To join the files, do the following:
cat codellama-70b-python-q6_K.gguf-split-* > codellama-70b-python-q6_K.gguf && rm codellama-70b-python-q6_K.gguf-split-*
Inference Providers
NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API:
The model has no library tag.