Yi-34B-200k-v2_GGUF / README.md
MarsupialAI's picture
Create README.md
62783ad verified
|
raw
history blame
No virus
641 Bytes
---
license: other
license_name: yi-license
license_link: LICENSE
---
GGUFs for Yi 34b 200k v2: https://huggingface.co/01-ai/Yi-34B-200K
This is the 24/03/07 updated version with enhanced long-context performance. From their notes:
> The long text capability of the Yi-34B-200K has been enhanced. In the "Needle-in-a-Haystack" test, the Yi-34B-200K's performance
is improved by 10.5%, rising from 89.3% to an impressive 99.8%. We continue to pretrain the model on 5B tokens long-context data
mixture and demonstrate a near-all-green performance.
FP16 split with peazip. Recombine with peazip, 7zip, or a simple concatenate command.