distilbert-base-uncased-fibe-v9-finetuned

This model is a fine-tuned version of pilotj/distilbert-base-uncased-fibe-v8-finetuned on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8968

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0001
  • train_batch_size: 32
  • eval_batch_size: 128
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 1

Training results

Training Loss Epoch Step Validation Loss
0.5704 0.0779 1000 1.0604
0.5614 0.1558 2000 1.1759
0.5481 0.2338 3000 1.0633
0.5417 0.3117 4000 1.0108
0.5257 0.3896 5000 0.9514
0.5106 0.4675 6000 0.9228
0.492 0.5455 7000 0.9493
0.5033 0.6234 8000 0.9060
0.4688 0.7013 9000 0.8691
0.483 0.7792 10000 0.8291
0.4679 0.8572 11000 0.8628
0.4368 0.9351 12000 0.8968

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0
  • Datasets 3.0.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
67M params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for pilotj/distilbert-base-uncased-fibe-v9-finetuned