--- license: apache-2.0 --- # gemma-7b-ultrachat-sft gemma-7b-ultrachat-sft is an SFT fine-tuned version of [google/gemma-7b](https://huggingface.co/google/gemma-7b) using the [stingning/ultrachat](https://huggingface.co/datasets/stingning/ultrachat) dataset. ## Fine-tuning configuration ### LoRA - LoRA r: 8 - LoRA alpha: 16 - LoRA dropout: 0.1 ### Training arguments - Epochs: 1 - Batch size: 4 - Gradient accumulation steps: 6 - Optimizer: paged_adamw_32bit - Max steps: 100 - Learning rate: 0.0002 - Weight decay: 0.001 - Learning rate scheduler type: constant - Max seq length: 2048