EVA Gen 0.0
Collection
RP/creative writing specialist models, trained on a curated mixture of natural and synthetic data.
•
5 items
•
Updated
•
1
An experimental online logit distillation of EVA-Qwen2.5-14B-v0.1 into Qwen2.5-1.5B. Should work as a RP/storywriting specialist, but don't expect superb performance from it, due to it's small size. All in all, it was a fun experiment to do.
Note: using quantized KV cache with Qwen2.5 is not recommended and can lead to degraded output quality. On the other hand, Qwen's KV cache is already light enough, so using f16 for it shouldn't be problematic.
Prompt format is ChatML.
Model was trained by Kearm and Auri.
Base model
Qwen/Qwen2.5-1.5B