|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- tay-yozhik/SyntheticTexts |
|
language: |
|
- ru |
|
--- |
|
# SyntheticRoBERTa |
|
|
|
This is a pre-trained model of type [RoBERTa](https://arxiv.org/abs/1907.11692). |
|
SyntheticRoBERTa is built on a dataset of texts in Russian, which were generated according to [the rules written in context-free grammar](https://github.com/GorbachevaTaisia/JSGF_generative_grammar). |
|
|
|
# Evaluation |
|
|
|
This model was evaluated on [RussianSuperGLUE tests](https://russiansuperglue.com/): |
|
| Task | Result | Metrics | |
|
|-------|----------|---------| |
|
| LiDiRus | 0,0 | Matthews Correlation Coefficient | |
|
| RCB | 0,091 / 0,158 | F1 / Accuracy | |
|
| PARus | 0,502 | Accuracy | |
|
| TERRa | 0,487 | Accuracy | |
|
| RUSSE | 0,587 | Accuracy | |
|
| RWSD | 0,331 | Accuracy | |