metadata
license: cc-by-nc-4.0
Command R+ GGUF
Description
This repository contains experimental GGUF weights that are currently compatible only with the following fork: https://github.com/Noeda/llama.cpp/tree/53f71f0026cbed4588b2ad16c51db630d2745794. I will update them once support for Command R+ is merged into the llama.cpp repository
Concatenating Weights
For every variant (except Q2_K), you must concatenate the weights, as they exceed the 50 GB single file size limit on HuggingFace. You can accomplish this using the cat
command on Linux (example for the Q3 variant):
cat command-r-plus-Q3_K_L-0000* > command-r-plus-Q3_K_L.gguf