|
--- |
|
language: |
|
- it |
|
pipeline_tag: fill-mask |
|
library_name: transformers |
|
widget: |
|
- text: "ciao" |
|
example_title: "ciao" |
|
--- |
|
|
|
This model (based on [bert-tiny](https://huggingface.co/google/bert_uncased_L-2_H-128_A-2)) has been trained from scratch on 20k italian Wikipedia articles and on a wide collection of italian words and dictionary definitions. |
|
|
|
The project is still a work in progress, new versions will come with time. |
|
|
|
# Training |
|
|
|
- epochs: 200 |
|
- lr: 1e-5 |
|
- optim: AdamW |
|
- weight_decay: 1e-3 |
|
|
|
# Eval |
|
|
|
- perplexity: 50 (it's a 12MB model, don't expect this to be ChatGPT anytime soon :) |