Zephyr 7B Gemma
Collection
Models, dataset, and Demo for Zephyr 7B Gemma. For code to train the models, see: https://github.com/huggingface/alignment-handbook
•
5 items
•
Updated
•
15
This model is a fine-tuned version of google/gemma-7b on the HuggingFaceH4/deita-10k-v0-sft dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss |
---|---|---|---|
0.9482 | 1.0 | 299 | 0.9848 |
0.8139 | 2.0 | 599 | 0.9610 |
0.722 | 2.99 | 897 | 0.9732 |
Base model
google/gemma-7b