--- license: agpl-3.0 language: - en pipeline_tag: text-generation --- # OpenChatRWKV 430m r2 This is a finetune of RWKV-v4neo 430m on `openchatgpt safe r2` dataset. `r2` shares no data with the `r1`, even the greetings, making this finetune, in some ways, inferiour to the original on `r1`. ## Key differences with `openchatrwkv-430m` One of the key differences is using an actual token for the instant message separation, apart from the new dataset. ## Differences with `openchatgpt-neox-125m` Increased parameter size and different dataset. ## Training data New dataset was obviously made at a later point in time. Many speculate that ChatGPT has degraded over the months, and I am a strong believer of that aswell - style in which model speaks started to sound different compared to 2-3 months ago. This model was trained on a mix of natural language and code. This model does not know how to greet you.