File size: 1,344 Bytes
cb20959 0420ce7 28deb88 892b20e db2d4cc b32d7c4 db2d4cc 892b20e 147ea7c fec3606 147ea7c fec3606 147ea7c fec3606 9c9fbd1 fec3606 892b20e 4e1b7fe d2a9f77 4e1b7fe 1b4c23c a76d32d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 |
---
license: apache-2.0
datasets:
- stingning/ultrachat
- kaist-ai/CoT-Collection
- mesolitica/google-translate-commitpackft
- Wanfq/Explore_Instruct_Rewriting_32k
- Wanfq/Explore_Instruct_Rewriting_10k
- Wanfq/Explore_Instruct_Brainstorming_16k
- xiyuez/red-dot-design-award-product-description
---
# RWKV v4 7B world model
finetuned with ultrachat , COT and some novel instructions data, commitpackft and so on
use full ultrachat and cot data, about 3B tokens
# Contributor
[@JL-er](https://huggingface.co/JL-er)
[@Remixa](https://huggingface.co/Remixa)
# Design of experiment
this model lose multi-turn chat ability,cause from using whole ultrachat datasets.
so i continue tuned multi-turn datasets with 2 aspects
1.role play
2.for novel multiturn instruction
# Training details
[wandb.ai](https://wandb.ai/one-/one-rwkv-64k)
# CAses
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/_1dJo549ldgX6q0JUwC6c.jpeg)
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/7969wbHaJpBq2n6xvfC7C.jpeg)
# Usage
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/cGDF6b4-x_9rcwMdl1KPp.png)
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6176b32847ee6431f632981e/hUxTVgjLBMcFqxQX9HoxL.png)
|