Effects of training data
#1
by
ChuckMcSneed
- opened
This model has a tendency to spit out stuff like <|prompt|> which was not present in base llama. It is also a bit drier than the base model. It seems that the training with Yarn has effects on output quality which were not observed in the benchmarks.