sberbank-ai
commited on
Commit
•
c5ba01c
1
Parent(s):
6777adb
Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ Architecture based on T5.
|
|
10 |
|
11 |
It has 24 layers and 1536 hidden size.
|
12 |
|
13 |
-
Model trained on a mixture of 7 denoisers like UL2 with several differences .
|
14 |
|
15 |
It trained on Russian language corpus (300GB). Dataset is the same as for ruT5 models.
|
16 |
|
|
|
10 |
|
11 |
It has 24 layers and 1536 hidden size.
|
12 |
|
13 |
+
Model trained on a mixture of 7 denoisers like UL2 with several differences (https://arxiv.org/abs/2205.05131).
|
14 |
|
15 |
It trained on Russian language corpus (300GB). Dataset is the same as for ruT5 models.
|
16 |
|