tburst commited on
Commit
c3b4bd1
1 Parent(s): 0b85278

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -26,7 +26,7 @@ It works for all languages the xlm-roberta model is pretrained on ([overview](ht
26
  ```python
27
  from transformers import AutoModelForSequenceClassification, AutoTokenizer
28
 
29
- model = AutoModelForSequenceClassification.from_pretrained("manifesto-project/manifestoberta-xlm-roberta-56policy-topics-sentence-2023-1")
30
  tokenizer = AutoTokenizer.from_pretrained("xlm-roberta-large")
31
 
32
  sentence = "We will restore funding to the Global Environment Facility and the Intergovernmental Panel on Climate Change, to support critical climate science research around the world"
@@ -60,5 +60,5 @@ The model was evaluated on a test set of 199,046 annotated manifesto statements.
60
 
61
  | | Accuracy | Top2_Acc | Top3_Acc | Precision| Recall | F1_Macro | MCC | Cross-Entropy |
62
  |-------------------------------------------------------------------------------------------------------|:--------:|:--------:|:--------:|:--------:|:------:|:--------:|:---:|:-------------:|
63
- [Sentence Model](https://huggingface.co/manifesto-project/xlm-roberta-political-56topics-sentence-2023a)| 0.57 | 0.73 | 0.81 | 0.49 | 0.43 | 0.45 | 0.55| 1.5 |
64
- [Context Model](https://huggingface.co/manifesto-project/xlm-roberta-political-56topics-context-2023a) | 0.64 | 0.81 | 0.88 | 0.54 | 0.52 | 0.53 | 0.62| 1.15 |
 
26
  ```python
27
  from transformers import AutoModelForSequenceClassification, AutoTokenizer
28
 
29
+ model = AutoModelForSequenceClassification.from_pretrained("manifesto-project/manifestoberta-xlm-roberta-56policy-topics-sentence-2023-1-1")
30
  tokenizer = AutoTokenizer.from_pretrained("xlm-roberta-large")
31
 
32
  sentence = "We will restore funding to the Global Environment Facility and the Intergovernmental Panel on Climate Change, to support critical climate science research around the world"
 
60
 
61
  | | Accuracy | Top2_Acc | Top3_Acc | Precision| Recall | F1_Macro | MCC | Cross-Entropy |
62
  |-------------------------------------------------------------------------------------------------------|:--------:|:--------:|:--------:|:--------:|:------:|:--------:|:---:|:-------------:|
63
+ [Sentence Model](https://huggingface.co/manifesto-project/manifestoberta-xlm-roberta-56policy-topics-sentence-2023-1-1)| 0.57 | 0.73 | 0.81 | 0.49 | 0.43 | 0.45 | 0.55| 1.5 |
64
+ [Context Model](https://huggingface.co/manifesto-project/manifestoberta-xlm-roberta-56policy-topics-context-2023-1-1) | 0.64 | 0.81 | 0.88 | 0.54 | 0.52 | 0.53 | 0.62| 1.15 |