distilroberta-base-finetuned-fake-news-english

This model is a fine-tuned version of distilroberta-base on the fake-and-real news dataset. It achieves the following results on the evaluation set:

  • Loss: 0.0020
  • Accuracy: 0.9997
  • F1: 0.9997
  • Precision: 0.9994
  • Recall: 1.0
  • Auc: 0.9997

Intended uses & limitations

The model may not work with the articles over 512 tokens after preprocessing as the model's context is restricted to a maximum of 512 tokens in the sequence.

Training and evaluation data

The fake-and-real news dataset contains a total of 44,898 annotated articles with 21,417 real and 23,481 fake. The dataset was stratified split into train, validation, and test subsets with a proportion of 60:20:20 respectively. The model was fine-tuned on the train subset and evaluated on validation and test subsets.

Split # examples
train 17959
validation 13469
test 13470

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 32
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 224
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Accuracy F1 Precision Recall Auc
0.251 0.36 200 0.0030 0.9996 0.9995 0.9995 0.9995 0.9996
0.0022 0.71 400 0.0012 0.9998 0.9998 0.9995 1.0 0.9998
0.0013 1.07 600 0.0001 1.0 1.0 1.0 1.0 1.0
0.0004 1.43 800 0.0015 0.9997 0.9997 0.9994 1.0 0.9997
0.0013 1.78 1000 0.0020 0.9997 0.9997 0.9994 1.0 0.9997

Framework versions

  • Transformers 4.17.0
  • Pytorch 1.10.0+cu111
  • Datasets 2.0.0
  • Tokenizers 0.12.0
Downloads last month
25
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.