Andrew Canis
commited on
Commit
•
01be733
1
Parent(s):
b605fd8
Update README
Browse filesMention how to compile the version of llama.cpp that works until the PR is merged upstream.
Also give command for verifying the md5sum
README.md
CHANGED
@@ -9,7 +9,20 @@ license: cc-by-nc-4.0
|
|
9 |
<!-- description start -->
|
10 |
## Description
|
11 |
|
12 |
-
This repo contains llama.cpp GGUF format model files for
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
## F16 files are split and require joining
|
15 |
|
@@ -36,3 +49,8 @@ Then you can remove the split files to save space:
|
|
36 |
del c4ai-command-r-v01-f16.gguf-split-a c4ai-command-r-v01-f16.gguf-split-b
|
37 |
```
|
38 |
|
|
|
|
|
|
|
|
|
|
|
|
9 |
<!-- description start -->
|
10 |
## Description
|
11 |
|
12 |
+
This repo contains llama.cpp GGUF format model files for
|
13 |
+
[Command-R 35B v1.0](https://huggingface.co/CohereForAI/c4ai-command-r-v01).
|
14 |
+
|
15 |
+
Note: you need to clone llama.cpp and compile until the
|
16 |
+
[PR6033](https://github.com/ggerganov/llama.cpp/pull/6033) is merged upstream:
|
17 |
+
```
|
18 |
+
git clone https://github.com/acanis/llama.cpp.git
|
19 |
+
cd llama.cpp
|
20 |
+
mkdir build
|
21 |
+
cd build
|
22 |
+
cmake .. -DLLAMA_CUBLAS=ON
|
23 |
+
cmake --build . --config Release -- -j16
|
24 |
+
cd ..
|
25 |
+
```
|
26 |
|
27 |
## F16 files are split and require joining
|
28 |
|
|
|
49 |
del c4ai-command-r-v01-f16.gguf-split-a c4ai-command-r-v01-f16.gguf-split-b
|
50 |
```
|
51 |
|
52 |
+
You can optionally confirm the checksum of merged c4ai-command-r-v01-f16.gguf
|
53 |
+
with the md5sum file:
|
54 |
+
```
|
55 |
+
md5sum -c md5sum
|
56 |
+
```
|