Gemma-2-2B
Collection
36 items
•
Updated
This model is a fine-tuned version of google/gemma-2-2b-it on the GaetanMichelet/chat-60_ft_task-1, the GaetanMichelet/chat-120_ft_task-1 and the GaetanMichelet/chat-180_ft_task-1 datasets. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
2.2103 | 1.0 | 17 | 2.0467 |
1.4945 | 2.0 | 34 | 1.5311 |
1.2419 | 3.0 | 51 | 1.3481 |
1.055 | 4.0 | 68 | 1.2944 |
0.9312 | 5.0 | 85 | 1.3499 |
0.6594 | 6.0 | 102 | 1.4708 |
0.3964 | 7.0 | 119 | 1.7099 |
0.2489 | 8.0 | 136 | 2.3134 |
0.1183 | 9.0 | 153 | 2.5791 |
0.0799 | 10.0 | 170 | 2.6094 |
0.0689 | 11.0 | 187 | 2.9233 |