lighteternal
commited on
Commit
•
dec7c0c
1
Parent(s):
f730ab2
Update README.md
Browse files
README.md
CHANGED
@@ -26,12 +26,15 @@ license: apache-2.0
|
|
26 |
|
27 |
|
28 |
This is an XLM-Roberta-base model with a binary classification head. Given a sentence, it can classify it either as a fact or an opinion based on its content.
|
29 |
-
You can use this model in any of the XLM-R supported languages for the same task, taking advantage of its 0-shot learning capabilities.
|
30 |
|
31 |
-
|
|
|
|
|
32 |
* Label 0: Opinion/Subjective sentence
|
33 |
* Label 1: Fact/Objective sentence
|
34 |
|
|
|
|
|
35 |
The original dataset (available here: https://github.com/1024er/cbert_aug/tree/crayon/datasets/subj) contained aprox. 9000 annotated sentences (classified as subjective or objective). It was translated to Greek using Google Translate. The Greek version was then concatenated with the original English one to create the mixed EN-EL dataset.
|
36 |
|
37 |
The model was trained for 5 epochs, using batch size = 8. Detailed metrics and hyperparameters available on the "Metrics" tab.
|
@@ -44,5 +47,6 @@ The model was trained for 5 epochs, using batch size = 8. Detailed metrics and h
|
|
44 |
|
45 |
|
46 |
## Acknowledgement
|
|
|
47 |
The research work was supported by the Hellenic Foundation for Research and Innovation (HFRI) under the HFRI PhD Fellowship grant (Fellowship Number:50, 2nd call)
|
48 |
|
|
|
26 |
|
27 |
|
28 |
This is an XLM-Roberta-base model with a binary classification head. Given a sentence, it can classify it either as a fact or an opinion based on its content.
|
|
|
29 |
|
30 |
+
You can use this model in any of the XLM-R supported languages for the same task, taking advantage of its 0-shot learning capabilities. However, the model was trained only using English and Greek sentences.
|
31 |
+
|
32 |
+
Legend of HuggingFace API labels:
|
33 |
* Label 0: Opinion/Subjective sentence
|
34 |
* Label 1: Fact/Objective sentence
|
35 |
|
36 |
+
## Dataset training info
|
37 |
+
|
38 |
The original dataset (available here: https://github.com/1024er/cbert_aug/tree/crayon/datasets/subj) contained aprox. 9000 annotated sentences (classified as subjective or objective). It was translated to Greek using Google Translate. The Greek version was then concatenated with the original English one to create the mixed EN-EL dataset.
|
39 |
|
40 |
The model was trained for 5 epochs, using batch size = 8. Detailed metrics and hyperparameters available on the "Metrics" tab.
|
|
|
47 |
|
48 |
|
49 |
## Acknowledgement
|
50 |
+
|
51 |
The research work was supported by the Hellenic Foundation for Research and Innovation (HFRI) under the HFRI PhD Fellowship grant (Fellowship Number:50, 2nd call)
|
52 |
|