bartowski commited on
Commit
601ae69
1 Parent(s): 6659e13

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -13,20 +13,22 @@ quantized_by: bartowski
13
 
14
  ## Llamacpp imatrix Quantizations of gemma-2-27b-it
15
 
16
- Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3277">b3277</a> for quantization.
17
 
18
  Original model: https://huggingface.co/google/gemma-2-27b-it
19
 
20
  All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
21
 
 
 
 
 
22
  ## Prompt format
23
 
24
  ```
25
- <bos><start_of_turn>user
26
  {prompt}<end_of_turn>
27
  <start_of_turn>model
28
- <end_of_turn>
29
- <start_of_turn>model
30
 
31
  ```
32
 
 
13
 
14
  ## Llamacpp imatrix Quantizations of gemma-2-27b-it
15
 
16
+ Using <a href="https://github.com/ggerganov/llama.cpp/">llama.cpp</a> release <a href="https://github.com/ggerganov/llama.cpp/releases/tag/b3278">b3278</a> for quantization.
17
 
18
  Original model: https://huggingface.co/google/gemma-2-27b-it
19
 
20
  All quants made using imatrix option with dataset from [here](https://gist.github.com/bartowski1182/eb213dccb3571f863da82e99418f81e8)
21
 
22
+ ## What's new
23
+
24
+ - July 21 2024: Contains latest tokenizer fixes, which addressed a few oddities from the original fix, should be closest to correct performance yet. Also has metadata for SWA and logit softcapping.
25
+
26
  ## Prompt format
27
 
28
  ```
29
+ <start_of_turn>user
30
  {prompt}<end_of_turn>
31
  <start_of_turn>model
 
 
32
 
33
  ```
34