--- license: llama2 --- Trained using TRL, it didn't fit properly on my 3090 without significantly dropping batch size and applying 4-bit quantization. It didn't exactly converge. ![training_run.png](https://cdn-uploads.huggingface.co/production/uploads/64075c834dc5f2846c96bc98/b-Tn5IDcRubZp_AyfLNg7.png)