Gemma-2-2B_auto
Collection
36 items
•
Updated
This model is a fine-tuned version of google/gemma-2-2b-it on the GaetanMichelet/chat-60_ft_task-3_auto, the GaetanMichelet/chat-120_ft_task-3_auto and the GaetanMichelet/chat-180_ft_task-3_auto datasets. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
1.3868 | 0.9412 | 8 | 1.3681 |
1.2861 | 2.0 | 17 | 1.2338 |
1.1878 | 2.9412 | 25 | 1.1353 |
0.9929 | 4.0 | 34 | 1.0232 |
0.9221 | 4.9412 | 42 | 0.9788 |
0.9315 | 6.0 | 51 | 0.9563 |
0.9202 | 6.9412 | 59 | 0.9455 |
0.8604 | 8.0 | 68 | 0.9398 |
0.7955 | 8.9412 | 76 | 0.9381 |
0.7864 | 10.0 | 85 | 0.9430 |
0.7442 | 10.9412 | 93 | 0.9520 |
0.6892 | 12.0 | 102 | 0.9797 |
0.6833 | 12.9412 | 110 | 0.9951 |
0.649 | 14.0 | 119 | 1.0528 |
0.5557 | 14.9412 | 127 | 1.1159 |
0.4908 | 16.0 | 136 | 1.1622 |