File size: 917 Bytes
eee565a f9ef747 eee565a f9ef747 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 |
---
license: agpl-3.0
language:
- en
pipeline_tag: text-generation
---
# OpenChatRWKV 430m r2
This is a finetune of RWKV-v4neo 430m on `openchatgpt safe r2` dataset. `r2` shares no data with the `r1`, even the greetings, making this finetune, in some ways, inferiour to the original on `r1`.
## Key differences with `openchatrwkv-430m`
One of the key differences is using an actual token for the instant message separation, apart from the new dataset.
## Differences with `openchatgpt-neox-125m`
Increased parameter size and different dataset.
## Training data
New dataset was obviously made at a later point in time. Many speculate that ChatGPT has degraded over the months, and I am a strong believer of that aswell - style in which model speaks started to sound different compared to 2-3 months ago.
This model was trained on a mix of natural language and code.
This model does not know how to greet you. |