Update README.md
Browse files
README.md
CHANGED
@@ -7,7 +7,7 @@
|
|
7 |
# Model
|
8 |
|
9 |
<!-- Provide a quick summary of what the model is/does. -->
|
10 |
-
This model card provides information about a fine-tuned T5 base model that has been specifically trained for generating summaries.
|
11 |
|
12 |
**Dataset and Training:**
|
13 |
The fine-tuned T5 base model is trained on a carefully curated subset of the XSum and ChatGPT datasets. These datasets contain a wide range of text samples, including news articles and conversational data. By utilizing this diverse data, the model gains a broader understanding of language and improves its ability to generate accurate and coherent summaries.
|
|
|
7 |
# Model
|
8 |
|
9 |
<!-- Provide a quick summary of what the model is/does. -->
|
10 |
+
This model card provides information about a fine-tuned T5 base model that has been specifically trained for generating summaries. We have made some key modifications to the training process to optimize the model's performance and provide the best possible summaries, particularly supporting greater length outputs. One notable difference between this model and other similar models is that it is trained on the target output length of 512. This means that the model is explicitly trained to generate summaries that are up to 512 tokens long. By focusing on this target output length, we aim to provide summaries that are more comprehensive and informative, while still maintaining a reasonable length for large text.
|
11 |
|
12 |
**Dataset and Training:**
|
13 |
The fine-tuned T5 base model is trained on a carefully curated subset of the XSum and ChatGPT datasets. These datasets contain a wide range of text samples, including news articles and conversational data. By utilizing this diverse data, the model gains a broader understanding of language and improves its ability to generate accurate and coherent summaries.
|