InferenceIllusionist
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -11,7 +11,7 @@ tags:
|
|
11 |
|
12 |
|
13 |
Quantized from fp32 with love. If you're using the latest version of llama.cpp you should no longer need to combine files before loading.
|
14 |
-
*
|
15 |
* Calculated in 105 chunks with n_ctx=512 using groups_merged.txt
|
16 |
|
17 |
For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
|
|
|
11 |
|
12 |
|
13 |
Quantized from fp32 with love. If you're using the latest version of llama.cpp you should no longer need to combine files before loading.
|
14 |
+
* Importance matrix calculated using fp16 precision model
|
15 |
* Calculated in 105 chunks with n_ctx=512 using groups_merged.txt
|
16 |
|
17 |
For a brief rundown of iMatrix quant performance please see this [PR](https://github.com/ggerganov/llama.cpp/pull/5747)
|