slm-experiments
Collection
15 items
•
Updated
This model is a fine-tuned version of dmariko/SmolLM-1.7B-Instruct_qlora_nf4_merged on the generator dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
No log | 0.8 | 1 | 1.7222 |
No log | 1.6 | 2 | 1.7181 |
No log | 2.4 | 3 | 1.6971 |
No log | 4.0 | 5 | 1.6586 |
No log | 4.8 | 6 | 1.6597 |
No log | 5.6 | 7 | 1.6572 |
No log | 6.4 | 8 | 1.6539 |
1.6809 | 8.0 | 10 | 1.6513 |
Base model
HuggingFaceTB/SmolLM-1.7B