ficsort's picture
Update README.md
83f1a20
|
raw
history blame
758 Bytes
---
language: hu
license: apache-2.0
datasets:
- wikipedia
tags:
- generated_from_keras_callback
- hubert
model-index:
- name: hubert-small-wiki
results: []
---
# hubert-small-wiki
This model was trained from scratch on the Wikipedia subset of Hungarian Webcorpus 2.0 with MLM and SOP tasks.
### Pre-Training Parameters:
First phase:
- Training steps: 500.000
- Sequence length: 128
- Batch size: 1024
Second phase:
- Training steps: 100.000
- Sequence length: 512
- Batch size: 384
### Framework versions
- Transformers 4.21.3
- TensorFlow 2.10.0
- Datasets 2.4.0
- Tokenizers 0.12.1
# Acknowledgement
[![Artificial Intelligence - National Laboratory - Hungary](https://milab.tk.hu/uploads/images/milab_logo_en.png)](https://mi.nemzetilabor.hu/)