eryk-mazus commited on
Commit
e2d9808
1 Parent(s): e9f2763

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -1
README.md CHANGED
@@ -43,4 +43,8 @@ Change `-ngl 32` to the number of layers to offload to GPU. Remove it if you don
43
 
44
  Change `-c 2048` to the desired sequence length. For extended sequence models - eg 8K, 16K, 32K - the necessary RoPE scaling parameters are read from the GGUF file and set by llama.cpp automatically.
45
 
46
- If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins`
 
 
 
 
 
43
 
44
  Change `-c 2048` to the desired sequence length. For extended sequence models - eg 8K, 16K, 32K - the necessary RoPE scaling parameters are read from the GGUF file and set by llama.cpp automatically.
45
 
46
+ If you want to have a chat-style conversation, replace the `-p <PROMPT>` argument with `-i -ins`
47
+
48
+ ## Notes on performance
49
+
50
+ For a model of this size, with stronger quantization, quality appears to decline much more than for larger models. Personally, I would advise to stick with `fp16` or `int8` for this model.