--- license: mit datasets: - pkupie/mc2_corpus language: - bo - ug - mn - kk --- # MC^2XLMR-large [Github Repo](https://github.com/luciusssss/mc2_corpus) We continually pretrain XLM-RoBERTa-large with [MC^2](https://huggingface.co/datasets/pkupie/mc2_corpus), which supports Tibetan, Uyghur, Kazakh in the Kazakh Arabic script, and Mongolian in the traditional Mongolian script. See details in the [paper](https://arxiv.org/abs/2311.08348). *We have also released another model trained on MC^2: [MC^2Llama-13B](https://huggingface.co/pkupie/mc2-llama-13b).* ## Citation ``` @article{zhang2024mc, title={MC$^2$: Towards Transparent and Culturally-Aware NLP for Minority Languages in China}, author={Zhang, Chen and Tao, Mingxu and Huang, Quzhe and Lin, Jiuheng and Chen, Zhibin and Feng, Yansong}, journal={arXiv preprint arXiv:2311.08348}, year={2024} } ```