File size: 1,920 Bytes
020f071 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
---
language:
- ru
- kbd
license: mit
base_model: facebook/m2m100_1.2B
tags:
- generated_from_trainer
datasets:
- anzorq/ru-kbd
model-index:
- name: m2m100_1.2B_ft_kbd-ru_63K
results: []
---
# m2m100_1.2B_ft_kbd-ru_63K
This model is a fine-tuned version of [facebook/m2m100_1.2B](https://huggingface.co/facebook/m2m100_1.2B) on the anzorq/ru-kbd dataset.
# Training Summary
- **Current Epoch**: 3.41/9
- **Global Step**: 9,000
- **Max Steps**: 23,778
- **Steps per Logging**: 500
- **Steps per Model Save**: 1,000
- **Total Operations (FLOPs)**: ~4.84 x 10^16
## Configuration:
- **Hyperparameter Search**: No
- **Is Local Process Zero**: True
- **Is World Process Zero**: True
## Progress:
| Epoch | Learning Rate | Loss | Step |
|-------|----------------------------------|--------|------|
| 0.19 | 4.8950710741021115e-05 | 2.4415 | 500 |
| 0.38 | 4.78993186979561e-05 | 1.7099 | 1000 |
| 0.57 | 4.684792665489108e-05 | 1.4997 | 1500 |
| 0.76 | 4.579653461182606e-05 | 1.3625 | 2000 |
| 0.95 | 4.4745142568761036e-05 | 1.2689 | 2500 |
| 1.14 | 4.3693750525696025e-05 | 1.0546 | 3000 |
| 1.32 | 4.264235848263101e-05 | 0.9711 | 3500 |
| 1.51 | 4.159096643956599e-05 | 0.9487 | 4000 |
| 1.70 | 4.05416771805871e-05 | 0.9202 | 4500 |
| 1.89 | 3.949238792160821e-05 | 0.8953 | 5000 |
| 2.08 | 3.844309866262932e-05 | 0.6436 | 5500 |
| 2.27 | 3.73917066195643e-05 | 0.6361 | 6000 |
| 2.46 | 3.6340314576499284e-05 | 0.6473 | 6500 |
| 2.65 | 3.52910253175204e-05 | 0.6383 | 7000 |
| 2.84 | 3.423963327445538e-05 | 0.6312 | 7500 |
| 3.03 | 3.318824123139036e-05 | 0.5965 | 8000 |
| 3.22 | 3.213684918832535e-05 | 0.4106 | 8500 |
| 3.41 | 3.1085457145260324e-05 | 0.4265 | 9000 | |