deberta-v3-large-AVeriTeC-nli
This model was finetuned from microsoft/deberta-v3-large on an AVeriTec dataset. It achieves the following results on the evaluation set:
Intended uses & limitations
This model is intended for usage in a pipeline for open-domain fact-checking task.
Training and evaluation data
See chenxwh/AVeriTeC
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- optimizer: adamw_torch
- training_precision: float16
- learning_rate: 1e-5
- per_device_train_batch_size: 32
- num_train_epochs: 10
- weight_decay: 0.01
- load_best_model_at_end: True #early stopping!
- warmup_ratio: 0.06
Training results
Framework versions
- Transformers 4.43.0
- TensorFlow 2.17.0
- Datasets 2.20.0
- Tokenizers 0.19.1
- Downloads last month
- 5
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.
Model tree for ctu-aic/deberta-v3-large-AVeriTeC-nli
Base model
microsoft/deberta-v3-largeEvaluation results
- dev macro F1 score on chenxwh/AVeriTeCself-reported0.710
- dev macro recall on chenxwh/AVeriTeCself-reported0.730
- dev macro precision on chenxwh/AVeriTeCself-reported0.710
- dev accuracy on chenxwh/AVeriTeCself-reported0.820