File size: 917 Bytes
eee565a
 
f9ef747
 
 
eee565a
f9ef747
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
---
license: agpl-3.0
language:
- en
pipeline_tag: text-generation
---

# OpenChatRWKV 430m r2

This is a finetune of RWKV-v4neo 430m on `openchatgpt safe r2` dataset. `r2` shares no data with the `r1`, even the greetings, making this finetune, in some ways, inferiour to the original on `r1`.

## Key differences with `openchatrwkv-430m`

One of the key differences is using an actual token for the instant message separation, apart from the new dataset.

## Differences with `openchatgpt-neox-125m`

Increased parameter size and different dataset.

## Training data

New dataset was obviously made at a later point in time. Many speculate that ChatGPT has degraded over the months, and I am a strong believer of that aswell - style in which model speaks started to sound different compared to 2-3 months ago.

This model was trained on a mix of natural language and code.

This model does not know how to greet you.