Any chance for a Gutenberg 9b from Gemma sppo?
I think this is one of the best finetunes for writing since a lot of the best writing models have this merged in at some point, and it also improved the model this was trained on in openllm average score if I remember right. The Gemma 9b sppo (iter 3) finetune I tested is surprisingly very good, and performs great in benchmarks (which is why I was skeptical and thought it would be bad or too overfit for real world use). I highly suggest giving dpo training it on the same dataset a try, preferably without any special tokens (the original Gemma 2 tokenizer should be preferred).
Sounds good! I'm working on a new finetuning setup, so this could be a good test for it.
Haven't tested it yet, but: https://huggingface.co/nbeerbower/gemma2-gutenberg-9B