Update README.md
Browse files
README.md
CHANGED
@@ -13,6 +13,8 @@ pipeline_tag: fill-mask
|
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
|
15 |
|
|
|
|
|
16 |
|
17 |
## Evaluation
|
18 |
|
@@ -32,6 +34,16 @@ Read the [paper](https://aclanthology.org/2024.unlp-1.14/) for more detailed tas
|
|
32 |
| [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
|
33 |
| [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
|
34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
|
37 |
## How to Get Started with the Model
|
|
|
13 |
<!-- Provide a quick summary of what the model is/does. -->
|
14 |
LiBERTa Large is a BERT-like model pre-trained from scratch exclusively for Ukrainian. It was presented during the [UNLP](https://unlp.org.ua/) @ [LREC-COLING 2024](https://lrec-coling-2024.org/). Further details are in the [LiBERTa: Advancing Ukrainian Language Modeling through Pre-training from Scratch](https://aclanthology.org/2024.unlp-1.14/) paper.
|
15 |
|
16 |
+
All the code is available in the [Goader/ukr-lm](https://github.com/Goader/ukr-lm) repository.
|
17 |
+
|
18 |
|
19 |
## Evaluation
|
20 |
|
|
|
34 |
| [liberta-large](https://huggingface.co/Goader/liberta-large) | 91.27 (1.22) | 92.50 (0.07) | 98.62 (0.08) | 95.44 (0.04) |
|
35 |
| [liberta-large-v2](https://huggingface.co/Goader/liberta-large-v2) | __91.73 (1.81)__ | __93.22 (0.14)__ | __98.79 (0.06)__ | 95.67 (0.12) |
|
36 |
|
37 |
+
## Fine-Tuning Hyperparameters
|
38 |
+
|
39 |
+
| Hyperparameter | Value |
|
40 |
+
|:---------------|:-----:|
|
41 |
+
| Peak Learning Rate | 3e-5 |
|
42 |
+
| Warm-up Ratio | 0.05 |
|
43 |
+
| Learning Rate Decay | Linear |
|
44 |
+
| Batch Size | 16 |
|
45 |
+
| Epochs | 10 |
|
46 |
+
| Weight Decay | 0.05 |
|
47 |
|
48 |
|
49 |
## How to Get Started with the Model
|