Edit model card

flan-t5-base-t5flan_finetune_reformat_in_given_manner

This model is a fine-tuned version of google/flan-t5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.0747
  • Rouge1: 80.2225
  • Rouge2: 78.7179
  • Rougel: 80.1485
  • Rougelsum: 80.0451
  • Gen Len: 17.3494

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 122 0.1377 79.3925 77.5379 79.3692 79.2301 17.1285
No log 2.0 244 0.1028 79.4291 77.762 79.3968 79.2743 17.1963
No log 3.0 366 0.0876 79.6022 77.9314 79.5449 79.4189 17.2004
No log 4.0 488 0.0808 79.9295 78.3351 79.8952 79.7778 17.2744
0.2046 5.0 610 0.0747 80.2225 78.7179 80.1485 80.0451 17.3494

Framework versions

  • Transformers 4.35.0
  • Pytorch 2.0.0
  • Datasets 2.15.0
  • Tokenizers 0.14.1
Downloads last month
2
Safetensors
Model size
248M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for ananttt/flan-t5-base-t5flan_finetune_reformat_in_given_manner

Finetuned
(621)
this model