zjyhf commited on
Commit
940cdc6
1 Parent(s): 1d88e17

The tokenizer adds a special token '<|im_end|>' to solve the problem of non-stop generation when encountering <|im_end|>.

Browse files

Using vllm to infer 'Llama3-ChatQA-1.5-70B', it will continue to be generated when encountering the special token '<|im_end|>', as shown in the figure below. This PR adds <|im_end|> to the tokenizer, and you need to add mapping to generation_config.json.
![8e4f01f676a0de25c1412b10172cfa9.png](https://cdn-uploads.huggingface.co/production/uploads/66161a077b605932bfbc106b/TSeY_EsLwBIZzh3LCMezu.png)

Files changed (1) hide show
  1. tokenizer.json +1 -1
tokenizer.json CHANGED
@@ -95,7 +95,7 @@
95
  },
96
  {
97
  "id": 128010,
98
- "content": "<|reserved_special_token_5|>",
99
  "single_word": false,
100
  "lstrip": false,
101
  "rstrip": false,
 
95
  },
96
  {
97
  "id": 128010,
98
+ "content": "<|im_end|>",
99
  "single_word": false,
100
  "lstrip": false,
101
  "rstrip": false,