skypro1111
commited on
Commit
•
4ac704b
1
Parent(s):
4799b36
Update README.md
Browse files
README.md
CHANGED
@@ -19,11 +19,11 @@ widget:
|
|
19 |
This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
|
20 |
|
21 |
## Training Data
|
22 |
-
The model was fine-tuned on a subset of
|
23 |
Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
|
24 |
|
25 |
## Training Procedure
|
26 |
-
The model underwent
|
27 |
|
28 |
```python
|
29 |
from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments
|
|
|
19 |
This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
|
20 |
|
21 |
## Training Data
|
22 |
+
The model was fine-tuned on a subset of 457,610 sentences from the Ubertext dataset, focusing on news content. The verbalized equivalents were created using Google Gemini Pro, providing a rich basis for learning text transformation tasks.
|
23 |
Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
|
24 |
|
25 |
## Training Procedure
|
26 |
+
The model underwent 410,000 training steps (1 epoch).
|
27 |
|
28 |
```python
|
29 |
from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments
|