jpodivin's picture
End of training
a0f0c9c verified
|
raw
history blame
2.34 kB
metadata
license: apache-2.0
base_model: distilbert-base-cased
tags:
  - generated_from_trainer
model-index:
  - name: distilbert-base-cased-logdetective-extraction-retrained
    results: []

distilbert-base-cased-logdetective-extraction-retrained

This model is a fine-tuned version of distilbert-base-cased on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 2.8926

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 20

Training results

Training Loss Epoch Step Validation Loss
No log 1.0 9 3.0408
No log 2.0 18 2.2720
No log 3.0 27 2.3287
No log 4.0 36 2.2797
No log 5.0 45 2.4768
No log 6.0 54 2.1528
No log 7.0 63 2.6022
No log 8.0 72 2.6710
No log 9.0 81 2.6752
No log 10.0 90 2.6804
No log 11.0 99 2.7833
No log 12.0 108 2.8071
No log 13.0 117 2.8256
No log 14.0 126 2.7876
No log 15.0 135 2.7295
No log 16.0 144 2.8091
No log 17.0 153 2.8284
No log 18.0 162 2.8335
No log 19.0 171 2.8810
No log 20.0 180 2.8926

Framework versions

  • Transformers 4.35.2
  • Pytorch 2.1.0+cu121
  • Datasets 2.16.1
  • Tokenizers 0.15.1