--- datasets: - samsum language: - en metrics: - rouge - bleu library_name: transformers pipeline_tag: summarization --- # t5-small-finetuned ## Model Description - **Purpose and Use**: This model is designed for abstractive text summarization with a focus on the SAMSum Dialogue Dataset. - **Model Architecture**: The architecture is based on a fine-tuned T5-small model, which consists of 60 million parameters. - **Training Data**: Trained on the SAMSum Dialogue Dataset, which comprises approximately 15,000 dialogue-summary pairs. ## Training Procedure - **Preprocessing**: Data preprocessing involved the removal of irrelevant tags and tokenization to ensure data consistency. - **Training Details**: The model was fine-tuned over 4 epochs with a learning rate of 2e-5 and a batch size of 2, utilizing gradient accumulation for optimization. - **Infrastructure**: Training was conducted using GPU acceleration and the Hugging Face Trainer API, with progress monitored via TensorBoard. ## Evaluation Results - **Metrics Used**: Evaluation metrics included ROUGE-1, ROUGE-2, ROUGE-L, BLEU, and Cosine Similarity. - **Performance**: The fine-tuned T5-small model demonstrated superior efficiency and effectiveness in summarization tasks, outperforming its larger counterparts. ## Validation and Test Set Performance | Metric | Validation Set | Test Set | |----------|--------------------|--------------| | ROUGE-1 | 0.5667 | 0.5536 | | ROUGE-2 | 0.2923 | 0.2718 | | ROUGE-L | 0.5306 | 0.5210 | The table above presents the performance of the model on both the validation and test sets, indicating the quality of content overlap and structural fluency in the summaries generated. ## Performance Metrics Comparison Across Models | Model | ROUGE-1 | ROUGE-2 | ROUGE-L | BLEU | Cosine Similarity | |----------|---------|---------|---------|------|-------------------| | My Model | 0.3767 | 0.1596 | 0.2896 | 9.52 | 0.7698 | | T5 Large | 0.3045 | 0.0960 | 0.2315 | 4.82 | 0.6745 | | Bart | 0.3189 | 0.0989 | 0.2352 | 6.28 | 0.6961 | | Pegasus | 0.2702 | 0.0703 | 0.2093 | 3.88 | 0.6432 | In the table above shows results on 50 samples for the test set that is being compared across various models.