Goader commited on
Commit
102ba64
1 Parent(s): 3d697bd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -0
README.md CHANGED
@@ -13,6 +13,8 @@ pipeline_tag: fill-mask
13
  <!-- Provide a quick summary of what the model is/does. -->
14
  LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
15
 
 
 
16
 
17
  ## Evaluation
18
 
@@ -32,6 +34,16 @@ Read the [paper](https://aclanthology.org/2024.unlp-1.14/) for more detailed tas
32
  | [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
33
  | [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
34
 
 
 
 
 
 
 
 
 
 
 
35
 
36
 
37
  ## How to Get Started with the Model
 
13
  <!-- Provide a quick summary of what the model is/does. -->
14
  LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
15
 
16
+ All the code is available in the [Goader/ukr-lm](https://github.com/Goader/ukr-lm) repository.
17
+
18
 
19
  ## Evaluation
20
 
 
34
  | [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
35
  | [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
36
 
37
+ ## Fine-Tuning Hyperparameters
38
+
39
+ | Hyperparameter | Value |
40
+ |:---------------|:-----:|
41
+ | Peak Learning Rate | 3e-5 |
42
+ | Warm-up Ratio | 0.05 |
43
+ | Learning Rate Decay | Linear |
44
+ | Batch Size | 16 |
45
+ | Epochs | 10 |
46
+ | Weight Decay | 0.05 |
47
 
48
 
49
  ## How to Get Started with the Model