deberta-small / README.md
lucadiliello's picture
Update README.md
fbf513c
metadata
datasets:
  - c4
language:
  - en
metrics:
  - accuracy
pipeline_tag: fill-mask

A small version of DeBERTa trained on the clean version of google C4 dataset. For more info about the size of the model, see config.json.

The model has been trained for 100K steps with a batch size of 2048 and a sequence length of 512, for a total of 104B tokens.

The vocabulary and the tokenizer are the same as microsoft/deberta-base.