indosum-lora-3 / README.md
apwic's picture
End of training
84184de verified
|
raw
history blame
2.01 kB
metadata
language:
  - id
license: apache-2.0
base_model: LazarusNLP/IndoNanoT5-base
tags:
  - generated_from_trainer
metrics:
  - rouge
model-index:
  - name: indosum-lora-3
    results: []

indosum-lora-3

This model is a fine-tuned version of LazarusNLP/IndoNanoT5-base on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5442
  • Rouge1: 72.6224
  • Rouge2: 65.7072
  • Rougel: 69.4293
  • Rougelsum: 71.6325
  • Gen Len: 101.7893

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.001
  • train_batch_size: 16
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5.0

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
0.8274 1.0 892 0.5684 69.5991 62.0811 66.2103 68.5179 100.152
0.6237 2.0 1784 0.5487 71.3379 64.0461 68.1399 70.3424 104.7813
0.5824 3.0 2676 0.5442 72.3075 65.3703 69.1638 71.3776 101.356
0.555 4.0 3568 0.5253 72.2139 65.2916 69.2278 71.2919 103.6467
0.5391 5.0 4460 0.5204 72.0194 64.9755 68.9244 71.0549 102.1173

Framework versions

  • Transformers 4.40.2
  • Pytorch 2.3.1+cu121
  • Datasets 2.20.0
  • Tokenizers 0.19.1