asier-gutierrez
commited on
Commit
•
099a437
1
Parent(s):
182de50
Update README.md
Browse files
README.md
CHANGED
@@ -33,7 +33,7 @@ Some of the statistics of the corpus:
|
|
33 |
The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [GPT-2](http://www.persagen.com/files/misc/radford2019language.pdf) model with a vocabulary size of 50,262 tokens. The GPT2-large-bne pre-training consists of an autoregressive language model training that follows the approach of the GPT-2. The training lasted a total of 10 days with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM.
|
34 |
|
35 |
## Evaluation and results
|
36 |
-
For evaluation details visit our [GitHub repository](https://github.com/PlanTL-
|
37 |
|
38 |
## Citing
|
39 |
Check out our paper for all the details: https://arxiv.org/abs/2107.07253
|
|
|
33 |
The training corpus has been tokenized using a byte version of Byte-Pair Encoding (BPE) used in the original [GPT-2](http://www.persagen.com/files/misc/radford2019language.pdf) model with a vocabulary size of 50,262 tokens. The GPT2-large-bne pre-training consists of an autoregressive language model training that follows the approach of the GPT-2. The training lasted a total of 10 days with 32 computing nodes each one with 4 NVIDIA V100 GPUs of 16GB VRAM.
|
34 |
|
35 |
## Evaluation and results
|
36 |
+
For evaluation details visit our [GitHub repository](https://github.com/PlanTL-GOB-ES/lm-spanish).
|
37 |
|
38 |
## Citing
|
39 |
Check out our paper for all the details: https://arxiv.org/abs/2107.07253
|