NaturalRoBERTa
This is a pre-trained model of type RoBERTa. NaturalRoBERTa is built on a dataset obtained from open sources: three news sub-corpuses Taiga (Lenta.ru, Interfax, N+1) and Russian Wikipedia texts.
Evaluation
This model was evaluated on RussianSuperGLUE tests:
Task | Result | Metrics |
---|---|---|
LiDiRus | 0,0 | Matthews Correlation Coefficient |
RCB | 0,217 / 0,484 | F1 / Accuracy |
PARus | 0,498 | Accuracy |
TERRa | 0,487 | Accuracy |
RUSSE | 0,587 | Accuracy |
RWSD | 0,669 | Accuracy |
- Downloads last month
- 4
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.