Update README.md
Browse files
README.md
CHANGED
@@ -35,6 +35,7 @@ widget:
|
|
35 |
- [Model Description](#model-description)
|
36 |
- [Intended Uses and Limitations](#intended-uses-and-limitations)
|
37 |
- [How to Use](#how-to-use)
|
|
|
38 |
- [Training](#training)
|
39 |
- [Training Data](#training-data)
|
40 |
- [Training Procedure](#training-procedure)
|
@@ -60,13 +61,9 @@ widget:
|
|
60 |
## Model Description
|
61 |
RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
|
62 |
|
63 |
-
|
64 |
## Intended Uses and Limitations
|
65 |
You can use the raw model for fill mask or fine-tune it to a downstream task.
|
66 |
|
67 |
-
The training data used for this model has not been released as a dataset one can browse. We know it contains a lot of
|
68 |
-
unfiltered content from the internet, which is far from neutral. At the time of submission, no measures have been taken to estimate the bias and toxicity embedded in the model. However, we are well aware that our models may be biased since the corpora have been collected using crawling techniques on multiple web sources. We intend to conduct research in these areas in the future, and if completed, this model card will be updated.
|
69 |
-
|
70 |
## How to Use
|
71 |
You can use this model directly with a pipeline for fill mask. Since the generation relies on some randomness, we set a seed for reproducibility:
|
72 |
|
@@ -109,6 +106,10 @@ Here is how to use this model to get the features of a given text in PyTorch:
|
|
109 |
torch.Size([1, 19, 1024])
|
110 |
```
|
111 |
|
|
|
|
|
|
|
|
|
112 |
## Training
|
113 |
|
114 |
### Training Data
|
|
|
35 |
- [Model Description](#model-description)
|
36 |
- [Intended Uses and Limitations](#intended-uses-and-limitations)
|
37 |
- [How to Use](#how-to-use)
|
38 |
+
- [Limitations and bias](#limitations-and-bias)
|
39 |
- [Training](#training)
|
40 |
- [Training Data](#training-data)
|
41 |
- [Training Procedure](#training-procedure)
|
|
|
61 |
## Model Description
|
62 |
RoBERTa-large-bne is a transformer-based masked language model for the Spanish language. It is based on the [RoBERTa](https://arxiv.org/abs/1907.11692) large model and has been pre-trained using the largest Spanish corpus known to date, with a total of 570GB of clean and deduplicated text processed for this work, compiled from the web crawlings performed by the [National Library of Spain (Biblioteca Nacional de España)](http://www.bne.es/en/Inicio/index.html) from 2009 to 2019.
|
63 |
|
|
|
64 |
## Intended Uses and Limitations
|
65 |
You can use the raw model for fill mask or fine-tune it to a downstream task.
|
66 |
|
|
|
|
|
|
|
67 |
## How to Use
|
68 |
You can use this model directly with a pipeline for fill mask. Since the generation relies on some randomness, we set a seed for reproducibility:
|
69 |
|
|
|
106 |
torch.Size([1, 19, 1024])
|
107 |
```
|
108 |
|
109 |
+
## Limitations and bias
|
110 |
+
|
111 |
+
At the time of submission, no measures have been taken to estimate the bias and toxicity embedded in the model. However, we are well aware that our models may be biased since the corpora have been collected using crawling techniques on multiple web sources. We intend to conduct research in these areas in the future, and if completed, this model card will be updated.
|
112 |
+
|
113 |
## Training
|
114 |
|
115 |
### Training Data
|