--- datasets: - bs-la/xP3ru license: bigscience-bloom-rail-1.0 --- # Model Summary [bloom-7b1](https://huggingface.co/bigscience/bloom-7b1) finetuned on Russian multitask data. Hence the same as [bloomz-7b1](https://huggingface.co/bigscience/bloomz-7b1), but with **only** Russian finetuning data. 500m stands for 500 million finetuning tokens. # Citation ``` BLOOM+1 - TODO ``` ```bibtex @misc{muennighoff2022crosslingual, title={Crosslingual Generalization through Multitask Finetuning}, author={Niklas Muennighoff and Thomas Wang and Lintang Sutawika and Adam Roberts and Stella Biderman and Teven Le Scao and M Saiful Bari and Sheng Shen and Zheng-Xin Yong and Hailey Schoelkopf and Xiangru Tang and Dragomir Radev and Alham Fikri Aji and Khalid Almubarak and Samuel Albanie and Zaid Alyafeai and Albert Webson and Edward Raff and Colin Raffel}, year={2022}, eprint={2211.01786}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```