--- language: - ru - kbd license: mit base_model: facebook/m2m100_1.2B tags: - generated_from_trainer datasets: - anzorq/ru-kbd model-index: - name: m2m100_1.2B_ft_kbd-ru_63K results: [] --- # m2m100_1.2B_ft_kbd-ru_63K This model is a fine-tuned version of [facebook/m2m100_1.2B](https://huggingface.co/facebook/m2m100_1.2B) on the anzorq/ru-kbd dataset. # Training Summary - **Current Epoch**: 3.41/9 - **Global Step**: 9,000 - **Max Steps**: 23,778 - **Steps per Logging**: 500 - **Steps per Model Save**: 1,000 - **Total Operations (FLOPs)**: ~4.84 x 10^16 ## Configuration: - **Hyperparameter Search**: No - **Is Local Process Zero**: True - **Is World Process Zero**: True ## Progress: | Epoch | Learning Rate | Loss | Step | |-------|----------------------------------|--------|------| | 0.19 | 4.8950710741021115e-05 | 2.4415 | 500 | | 0.38 | 4.78993186979561e-05 | 1.7099 | 1000 | | 0.57 | 4.684792665489108e-05 | 1.4997 | 1500 | | 0.76 | 4.579653461182606e-05 | 1.3625 | 2000 | | 0.95 | 4.4745142568761036e-05 | 1.2689 | 2500 | | 1.14 | 4.3693750525696025e-05 | 1.0546 | 3000 | | 1.32 | 4.264235848263101e-05 | 0.9711 | 3500 | | 1.51 | 4.159096643956599e-05 | 0.9487 | 4000 | | 1.70 | 4.05416771805871e-05 | 0.9202 | 4500 | | 1.89 | 3.949238792160821e-05 | 0.8953 | 5000 | | 2.08 | 3.844309866262932e-05 | 0.6436 | 5500 | | 2.27 | 3.73917066195643e-05 | 0.6361 | 6000 | | 2.46 | 3.6340314576499284e-05 | 0.6473 | 6500 | | 2.65 | 3.52910253175204e-05 | 0.6383 | 7000 | | 2.84 | 3.423963327445538e-05 | 0.6312 | 7500 | | 3.03 | 3.318824123139036e-05 | 0.5965 | 8000 | | 3.22 | 3.213684918832535e-05 | 0.4106 | 8500 | | 3.41 | 3.1085457145260324e-05 | 0.4265 | 9000 |