Commit
•
3adecfd
1
Parent(s):
1fcdf56
Update README.md
Browse files
README.md
CHANGED
@@ -100,7 +100,7 @@ model-index: # info: https://github.com/huggingface/hub-docs/blame/main/modelca
|
|
100 |
|
101 |
# DeBERTa-v3-large-mnli-fever-anli-ling-wanli
|
102 |
## Model description
|
103 |
-
This model was fine-tuned on the [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), Adversarial-NLI ([ANLI](https://huggingface.co/datasets/anli)), [LingNLI](https://arxiv.org/pdf/2104.07179.pdf) and [WANLI](https://huggingface.co/datasets/alisawuffles/WANLI) datasets, which comprise 885 242 NLI hypothesis-premise pairs. This model is the best NLI
|
104 |
|
105 |
The foundation model is [DeBERTa-v3-large from Microsoft](https://huggingface.co/microsoft/deberta-v3-large). Released on 06.12.21, DeBERTa-v3-large is currently the best large-sized foundation model for text classification. It combines several recent innovations compared to classical Masked Language Models like BERT, RoBERTa etc., see the [paper](https://arxiv.org/abs/2111.09543)
|
106 |
|
|
|
100 |
|
101 |
# DeBERTa-v3-large-mnli-fever-anli-ling-wanli
|
102 |
## Model description
|
103 |
+
This model was fine-tuned on the [MultiNLI](https://huggingface.co/datasets/multi_nli), [Fever-NLI](https://github.com/easonnie/combine-FEVER-NSMN/blob/master/other_resources/nli_fever.md), Adversarial-NLI ([ANLI](https://huggingface.co/datasets/anli)), [LingNLI](https://arxiv.org/pdf/2104.07179.pdf) and [WANLI](https://huggingface.co/datasets/alisawuffles/WANLI) datasets, which comprise 885 242 NLI hypothesis-premise pairs. This model is the best performing NLI model on the Hugging Face Hub as of 06.06.22 and can be used for zero-shot classification. It significantly outperforms all other large models on the [ANLI benchmark](https://github.com/facebookresearch/anli).
|
104 |
|
105 |
The foundation model is [DeBERTa-v3-large from Microsoft](https://huggingface.co/microsoft/deberta-v3-large). Released on 06.12.21, DeBERTa-v3-large is currently the best large-sized foundation model for text classification. It combines several recent innovations compared to classical Masked Language Models like BERT, RoBERTa etc., see the [paper](https://arxiv.org/abs/2111.09543)
|
106 |
|