saattrupdan
commited on
Commit
•
c26025b
1
Parent(s):
016baab
Update README.md
Browse files
README.md
CHANGED
@@ -32,9 +32,13 @@ inference:
|
|
32 |
|
33 |
This model is a fine-tuned version of [NbAiLab/nb-bert-large](https://huggingface.co/NbAiLab/nb-bert-large) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
|
34 |
|
35 |
-
|
36 |
|
37 |
-
|
|
|
|
|
|
|
|
|
38 |
|
39 |
|
40 |
## Quick start
|
@@ -68,13 +72,17 @@ We report Matthew's Correlation Coefficient (MCC), macro-average F1-score as wel
|
|
68 |
|
69 |
| **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
|
70 |
| :-------- | :------------ | :--------- | :----------- | :----------- |
|
71 |
-
| `alexandrainst/
|
72 |
-
| [`alexandrainst/
|
73 |
-
| [`alexandrainst/
|
74 |
|
75 |
|
76 |
## Training procedure
|
77 |
|
|
|
|
|
|
|
|
|
78 |
### Training hyperparameters
|
79 |
|
80 |
The following hyperparameters were used during training:
|
|
|
32 |
|
33 |
This model is a fine-tuned version of [NbAiLab/nb-bert-large](https://huggingface.co/NbAiLab/nb-bert-large) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
|
34 |
|
35 |
+
We have released three models for Scandinavian NLI, of different sizes:
|
36 |
|
37 |
+
- [alexandrainst/scandi-nli-large](https://huggingface.co/alexandrainst/scandi-nli-large)
|
38 |
+
- [alexandrainst/scandi-nli-base](https://huggingface.co/alexandrainst/scandi-nli-base)
|
39 |
+
- [alexandrainst/scandi-nli-small](https://huggingface.co/alexandrainst/scandi-nli-small)
|
40 |
+
|
41 |
+
The performance and model size of each of them can be found in the [Performance]() section below.
|
42 |
|
43 |
|
44 |
## Quick start
|
|
|
72 |
|
73 |
| **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
|
74 |
| :-------- | :------------ | :--------- | :----------- | :----------- |
|
75 |
+
| `alexandrainst/scandi-nli-large` (this) | **73.80%** | **58.41%** | **86.98%** | 354M |
|
76 |
+
| [`alexandrainst/scandi-nli-base`](https://huggingface.co/alexandrainst/scandi-nli-base) | 62.44% | 55.00% | 80.42% | 178M |
|
77 |
+
| [`alexandrainst/scandi-nli-small`](https://huggingface.co/alexandrainst/scandi-nli-small) | 47.28% | 48.88% | 73.46% | **22M** |
|
78 |
|
79 |
|
80 |
## Training procedure
|
81 |
|
82 |
+
It has been fine-tuned on a dataset composed of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) as well as machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) and [CommitmentBank](https://doi.org/10.18148/sub/2019.v23i2.601) into all three languages, and machine translated versions of [FEVER](https://aclanthology.org/N18-1074/) and [Adversarial NLI](https://aclanthology.org/2020.acl-main.441/) into Swedish.
|
83 |
+
|
84 |
+
The three languages are sampled equally during training, and they're validated on validation splits of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) and machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) for Swedish and Norwegian Bokmål, sampled equally.
|
85 |
+
|
86 |
### Training hyperparameters
|
87 |
|
88 |
The following hyperparameters were used during training:
|