dranger003 commited on
Commit
4f67f3f
1 Parent(s): 17680aa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -12,11 +12,11 @@ The quants here are meant to test imatrix quantized weights.
12
 
13
  **Added `ggml-dbrx-instruct-16x12b-f16_imatrix-wiki.dat` which is a 2K batches (1M tokens) on FP16 weights using wiki.train.**
14
 
15
- | Precision | Quant/imatrix | Size (GiB) | PPL |
16
  | -- | -- | -- | -- |
17
- | IQ4_XS | Q8_0/wiki.train | 65.29 | 5.2260 +/- 0.03558 |
18
- | IQ4_XS | FP16/wiki.train | 65.29 | 5.2241 +/- 0.03559 |
19
- | IQ4_XS | None | 66.05 | 5.2546 +/- 0.03570 |
20
 
21
  **2024-04-13**: Support for this model has just being merged - [`PR #6515`](https://github.com/ggerganov/llama.cpp/pull/6515).
22
  **<u>You will need this llama.cpp commit [`4bd0f93e`](https://github.com/ggerganov/llama.cpp/commit/4bd0f93e4ab4fe6682e7d0241c1bdec1397e954a) to run this model</u>**
 
12
 
13
  **Added `ggml-dbrx-instruct-16x12b-f16_imatrix-wiki.dat` which is a 2K batches (1M tokens) on FP16 weights using wiki.train.**
14
 
15
+ | Quant | IMatrix Quant/Dataset/Chunks | Size (GiB) | PPL (wiki.test) |
16
  | -- | -- | -- | -- |
17
+ | IQ4_XS | Q8_0/wiki.train/200 | 65.29 | 5.2260 +/- 0.03558 |
18
+ | IQ4_XS | FP16/wiki.train/2000 | 65.29 | 5.2241 +/- 0.03559 |
19
+ | IQ4_XS | - | 66.05 | 5.2546 +/- 0.03570 |
20
 
21
  **2024-04-13**: Support for this model has just being merged - [`PR #6515`](https://github.com/ggerganov/llama.cpp/pull/6515).
22
  **<u>You will need this llama.cpp commit [`4bd0f93e`](https://github.com/ggerganov/llama.cpp/commit/4bd0f93e4ab4fe6682e7d0241c1bdec1397e954a) to run this model</u>**