|
--- |
|
language: |
|
- ar |
|
|
|
datasets: |
|
- labr |
|
|
|
tags: |
|
- labr |
|
|
|
widget: |
|
- text: "كتاب ممل جدا تضييع وقت" |
|
- text: "اسلوب ممتع وشيق في الكتاب استمعت بالاحداث" |
|
|
|
--- |
|
|
|
# BERT-LABR unbalanced |
|
Arabic version bert model fine tuned on LABR dataset |
|
|
|
## Data |
|
The model were fine-tuned on ~63000 book reviews in arabic using bert large arabic |
|
|
|
|
|
## Results |
|
| class | precision | recall | f1-score | Support | |
|
|----------|-----------|--------|----------|---------| |
|
| 0 | 0.8109 | 0.6832 | 0.7416 | 1670 | |
|
| 1 | 0.9399 | 0.9689 | 0.9542 | 8541 | |
|
| Accuracy | | | 0.9221 | 10211 | |
|
|
|
|
|
|
|
## How to use |
|
|
|
You can use these models by installing `torch` or `tensorflow` and Huggingface library `transformers`. And you can use it directly by initializing it like this: |
|
|
|
```python |
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer |
|
|
|
model_name="mofawzy/bert-labr-unbalanced" |
|
model = AutoModelForSequenceClassification.from_pretrained(model_name,num_labels=2) |
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
|
``` |
|
|