Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,7 @@ license: mit
|
|
9 |
|
10 |
Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates.
|
11 |
|
12 |
-
This the DeBERTa V2 xxlarge model
|
13 |
|
14 |
|
15 |
#### Fine-tuning on NLU tasks
|
@@ -24,7 +24,7 @@ We present the dev results on SQuAD 1.1/2.0 and several GLUE benchmark tasks.
|
|
24 |
| DeBERTa-Large | 95.5/90.1 | 90.7/88.0 | 91.3/91.1 | 96.5 | 95.3 | 69.5 | 86.6 | 92.6/94.6 | 92.3 |92.5 |
|
25 |
| DeBERTa-XLarge | -/- | -/- | 91.5/91.0 | - | - | - | 89.5 | 92.1/94.3 | - |- |
|
26 |
| DeBERTa-XLarge-V2 | - | - | 91.7/91.6 | - | - | - | - | - | - |- |
|
27 |
-
|**DeBERTa-XXLarge-V2
|
28 |
| DeBERTa-XLarge-V2-mnli | - | - | 91.7/91.6 | - | - | - | 93.9 | - | - |- |
|
29 |
|**DeBERTa-XXLarge-V2-mnli**| - | - |**91.7/91.8**| - | - | - | 93.5 | - | - |- |
|
30 |
|
|
|
9 |
|
10 |
Please check the [official repository](https://github.com/microsoft/DeBERTa) for more details and updates.
|
11 |
|
12 |
+
This is the DeBERTa V2 xxlarge model with 48 layers, 1536 hidden size. Total parameters 1.5B. It's trained with 160GB data.
|
13 |
|
14 |
|
15 |
#### Fine-tuning on NLU tasks
|
|
|
24 |
| DeBERTa-Large | 95.5/90.1 | 90.7/88.0 | 91.3/91.1 | 96.5 | 95.3 | 69.5 | 86.6 | 92.6/94.6 | 92.3 |92.5 |
|
25 |
| DeBERTa-XLarge | -/- | -/- | 91.5/91.0 | - | - | - | 89.5 | 92.1/94.3 | - |- |
|
26 |
| DeBERTa-XLarge-V2 | - | - | 91.7/91.6 | - | - | - | - | - | - |- |
|
27 |
+
|**DeBERTa-XXLarge-V2**| 96.1/91.4 | 92.2/89.7 |**91.7/91.8**| - | - | - | - | - | - |- |
|
28 |
| DeBERTa-XLarge-V2-mnli | - | - | 91.7/91.6 | - | - | - | 93.9 | - | - |- |
|
29 |
|**DeBERTa-XXLarge-V2-mnli**| - | - |**91.7/91.8**| - | - | - | 93.5 | - | - |- |
|
30 |
|