Initial GGUF model commit
Browse files
README.md
CHANGED
@@ -204,6 +204,10 @@ This is Transformers/HF format fp16 weights for CodeLlama 7B-Python. It is the
|
|
204 |
|
205 |
Quantisations will be coming shortly.
|
206 |
|
|
|
|
|
|
|
|
|
207 |
## Prompt template: TBC
|
208 |
|
209 |
|
|
|
204 |
|
205 |
Quantisations will be coming shortly.
|
206 |
|
207 |
+
Please note that due to a change in the RoPE Theta value, for correct results you must load these FP16 models with `trust_remote_code=True`
|
208 |
+
|
209 |
+
Credit to @emozilla for creating the necessary modelling code to achieve this!
|
210 |
+
|
211 |
## Prompt template: TBC
|
212 |
|
213 |
|