skypro1111 commited on
Commit
bc487eb
1 Parent(s): 698b693

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -18,11 +18,11 @@ widget:
18
  This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
19
 
20
  ## Training Data
21
- The model was fine-tuned on a subset of 96,780 sentences from the Ubertext dataset, focusing on news content. The verbalized equivalents were created using Google Gemini Pro, providing a rich basis for learning text transformation tasks.
22
  Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
23
 
24
  ## Training Procedure
25
- The model underwent 70,000 training steps, which is almost 2 epochs, with further training the results degraded.
26
 
27
  ```python
28
  from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments
 
18
  This model is based on the [facebook/mbart-large-50](https://huggingface.co/facebook/mbart-large-50) architecture, renowned for its effectiveness in translation and text generation tasks across numerous languages.
19
 
20
  ## Training Data
21
+ The model was fine-tuned on a subset of 457,610 sentences from the Ubertext dataset, focusing on news content. The verbalized equivalents were created using Google Gemini Pro, providing a rich basis for learning text transformation tasks.
22
  Dataset [skypro1111/ubertext-2-news-verbalized](https://huggingface.co/datasets/skypro1111/ubertext-2-news-verbalized)
23
 
24
  ## Training Procedure
25
+ The model underwent 410,000 training steps.
26
 
27
  ```python
28
  from transformers import MBartForConditionalGeneration, AutoTokenizer, Trainer, TrainingArguments