xiaol's picture
Update README.md
9c9fbd1
|
raw
history blame
1.34 kB
metadata
license: apache-2.0
datasets:
  - stingning/ultrachat
  - kaist-ai/CoT-Collection
  - mesolitica/google-translate-commitpackft
  - Wanfq/Explore_Instruct_Rewriting_32k
  - Wanfq/Explore_Instruct_Rewriting_10k
  - Wanfq/Explore_Instruct_Brainstorming_16k
  - xiyuez/red-dot-design-award-product-description

RWKV v4 7B world model

finetuned with ultrachat , COT and some novel instructions data, commitpackft and so on

use full ultrachat and cot data, about 3B tokens

Contributor

@JL-er @Remixa

Design of experiment

this model lose multi-turn chat ability,cause from using whole ultrachat datasets.

so i continue tuned multi-turn datasets with 2 aspects

1.role play

2.for novel multiturn instruction

Training details

wandb.ai

CAses

image/jpeg

image/jpeg

Usage

image/png

image/png