EOS Token

#1
by ssmi153 - opened

I'm finding that these quantized models don't know how to stop. Comparing your special_tokens_map.json to the original repo, you have a different "eos_token" value. I think this should be "<|end_of_turn|>" rather than "</s>".

Ah yeah, they must have edited their JSON after I did my quantisations. I have edited my JSON to match theirs, so do a re-download of those files and hopefully that will sort it out.

Could it be that TheBloke/vicuna-13B-v1.5-16K-GPTQ suffers from the same problem?

Could it be that TheBloke/vicuna-13B-v1.5-16K-GPTQ suffers from the same problem?

Don't think so - Vicuna has always used </s> and this hasn't changed

Thanks for the answer! Your newer GPTQ-models don't know when to stop in 1 of five cases on my end. The original vicuna-13B-v1.5-16K seems to be fine though..

Sign up or log in to comment