Zero-Shot Classification
Transformers
PyTorch
Safetensors
bert
text-classification
Inference Endpoints
saattrupdan commited on
Commit
c26025b
1 Parent(s): 016baab

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -5
README.md CHANGED
@@ -32,9 +32,13 @@ inference:
32
 
33
  This model is a fine-tuned version of [NbAiLab/nb-bert-large](https://huggingface.co/NbAiLab/nb-bert-large) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
34
 
35
- It has been fine-tuned on a dataset composed of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) as well as machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) and [CommitmentBank](https://doi.org/10.18148/sub/2019.v23i2.601) into all three languages, and machine translated versions of [FEVER](https://aclanthology.org/N18-1074/) and [Adversarial NLI](https://aclanthology.org/2020.acl-main.441/) into Swedish.
36
 
37
- The three languages are sampled equally during training, and they're validated on validation splits of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) and machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) for Swedish and Norwegian Bokmål, sampled equally.
 
 
 
 
38
 
39
 
40
  ## Quick start
@@ -68,13 +72,17 @@ We report Matthew's Correlation Coefficient (MCC), macro-average F1-score as wel
68
 
69
  | **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
70
  | :-------- | :------------ | :--------- | :----------- | :----------- |
71
- | `alexandrainst/nb-bert-large-nli-scandi` (this) | **73.80%** | **58.41%** | **86.98%** | 354M |
72
- | [`alexandrainst/nb-bert-base-nli-scandi`](https://huggingface.co/alexandrainst/nb-bert-base-nli-scandi) | 62.44% | 55.00% | 80.42% | 178M |
73
- | [`alexandrainst/electra-small-nordic-nli-scandi`](https://huggingface.co/alexandrainst/electra-small-nordic-nli-scandi) | 47.28% | 48.88% | 73.46% | **22M** |
74
 
75
 
76
  ## Training procedure
77
 
 
 
 
 
78
  ### Training hyperparameters
79
 
80
  The following hyperparameters were used during training:
 
32
 
33
  This model is a fine-tuned version of [NbAiLab/nb-bert-large](https://huggingface.co/NbAiLab/nb-bert-large) for Natural Language Inference in Danish, Norwegian Bokmål and Swedish.
34
 
35
+ We have released three models for Scandinavian NLI, of different sizes:
36
 
37
+ - [alexandrainst/scandi-nli-large](https://huggingface.co/alexandrainst/scandi-nli-large)
38
+ - [alexandrainst/scandi-nli-base](https://huggingface.co/alexandrainst/scandi-nli-base)
39
+ - [alexandrainst/scandi-nli-small](https://huggingface.co/alexandrainst/scandi-nli-small)
40
+
41
+ The performance and model size of each of them can be found in the [Performance]() section below.
42
 
43
 
44
  ## Quick start
 
72
 
73
  | **Model** | **MCC** | **Macro-F1** | **Accuracy** | **Number of Parameters** |
74
  | :-------- | :------------ | :--------- | :----------- | :----------- |
75
+ | `alexandrainst/scandi-nli-large` (this) | **73.80%** | **58.41%** | **86.98%** | 354M |
76
+ | [`alexandrainst/scandi-nli-base`](https://huggingface.co/alexandrainst/scandi-nli-base) | 62.44% | 55.00% | 80.42% | 178M |
77
+ | [`alexandrainst/scandi-nli-small`](https://huggingface.co/alexandrainst/scandi-nli-small) | 47.28% | 48.88% | 73.46% | **22M** |
78
 
79
 
80
  ## Training procedure
81
 
82
+ It has been fine-tuned on a dataset composed of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) as well as machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) and [CommitmentBank](https://doi.org/10.18148/sub/2019.v23i2.601) into all three languages, and machine translated versions of [FEVER](https://aclanthology.org/N18-1074/) and [Adversarial NLI](https://aclanthology.org/2020.acl-main.441/) into Swedish.
83
+
84
+ The three languages are sampled equally during training, and they're validated on validation splits of [DanFEVER](https://aclanthology.org/2021.nodalida-main.pdf#page=439) and machine translated versions of [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) for Swedish and Norwegian Bokmål, sampled equally.
85
+
86
  ### Training hyperparameters
87
 
88
  The following hyperparameters were used during training: