dantedgp's picture
End of training
64bce66 verified
|
raw
history blame
2.41 kB
metadata
license: apache-2.0
base_model: google/flan-t5-small
tags:
  - generated_from_trainer
metrics:
  - rouge
model-index:
  - name: flan-t5-small-finetuned-question-generation
    results: []

flan-t5-small-finetuned-question-generation

This model is a fine-tuned version of google/flan-t5-small on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 1.5888
  • Rouge1: 49.9368
  • Rouge2: 27.2185
  • Rougel: 46.1729
  • Rougelsum: 46.1900
  • Gen Len: 13.7801

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5.6e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 8

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.8453 1.0 10917 1.6093 47.9986 25.3028 44.3072 44.3345 13.6285
1.6998 2.0 21834 1.5955 49.1185 26.2998 45.4201 45.4194 13.4627
1.6257 3.0 32751 1.5933 49.4151 26.7655 45.6769 45.6809 13.7603
1.5722 4.0 43668 1.5822 49.4727 26.8015 45.8010 45.8247 13.6301
1.5337 5.0 54585 1.5881 49.5572 26.8461 45.8811 45.8919 13.7029
1.5045 6.0 65502 1.5852 49.7863 27.0921 46.0625 46.0677 13.8099
1.4822 7.0 76419 1.5867 49.7769 27.0550 46.0091 46.0228 13.8027
1.4674 8.0 87336 1.5888 49.9368 27.2185 46.1729 46.1900 13.7801

Framework versions

  • Transformers 4.42.4
  • Pytorch 2.3.1
  • Datasets 2.20.0
  • Tokenizers 0.19.1