dumitrescustefan
commited on
Commit
•
a14be92
1
Parent(s):
23d4ced
Update README.md
Browse files
README.md
CHANGED
@@ -4,9 +4,7 @@ inference: false
|
|
4 |
license: apache-2.0
|
5 |
---
|
6 |
|
7 |
-
|
8 |
-
|
9 |
-
This is a pretrained [MT5](https://github.com/google-research/multilingual-t5) large model (973M parameters).
|
10 |
|
11 |
Training was performed with the span corruption task on a clean 80GB Romanian text corpus for 4M total steps with these [scripts](https://github.com/dumitrescustefan/t5x_models), starting from the 1M public mt5x-large checkpoint. The model was trained with an encoder and decoder sequence length of 512, and has the same mt5x vocabulary as the 1M multilingual checkpoint.
|
12 |
|
|
|
4 |
license: apache-2.0
|
5 |
---
|
6 |
|
7 |
+
This is a pretrained [MT5](https://github.com/google-research/multilingual-t5) large model (**973M** parameters).
|
|
|
|
|
8 |
|
9 |
Training was performed with the span corruption task on a clean 80GB Romanian text corpus for 4M total steps with these [scripts](https://github.com/dumitrescustefan/t5x_models), starting from the 1M public mt5x-large checkpoint. The model was trained with an encoder and decoder sequence length of 512, and has the same mt5x vocabulary as the 1M multilingual checkpoint.
|
10 |
|