tastypear commited on
Commit
c08e2d1
1 Parent(s): b7e036d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -0
README.md CHANGED
@@ -1,3 +1,38 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ Original model: [xiaol/RWKV-v5-12B-one-state-chat-16k](https://huggingface.co/xiaol/RWKV-v5-12B-one-state-chat-16k)
6
+
7
+ You can run this model with [ai00_rwkv_server](https://github.com/cgisky1980/ai00_rwkv_server).
8
+
9
+ Although ai00_rwkv_server is mainly for lowend PC, you can run it on servers which are support VULKAN.
10
+
11
+ To try it in Colab, you should install [libnvidia-gl-*](https://packages.ubuntu.com/search?keywords=libnvidia-gl&searchon=names&suite=jammy&section=all) :
12
+ ```python
13
+ !apt -y install libnvidia-gl-535
14
+ ```
15
+
16
+ ----
17
+
18
+ # Original model card:
19
+
20
+ # Release date: December 18th
21
+
22
+ finetuned from the state-of-the-art (SOTA) model, RWKV v5 12B one state base! More details will be provided soon. Enjoy the incredible performance of this model, which is optimized for systems with 24GB of VRAM and supports fp16. It can be fine-tuned using a single A100 GPU. To execute this model, utilize the [RWKV Runner](https://github.com/josStorer/RWKV-Runner) tool.
23
+
24
+ # Finetuned from [Mobius 12B base](https://huggingface.co/xiaol/Mobius-12B-base)
25
+
26
+ # Usage
27
+ - [RWKV next web](https://rwkv-next-web.ai-creator.net/)
28
+ - if use with [RWKV runner](https://github.com/josStorer/RWKV-Runner) or [ai00 server](https://github.com/cgisky1980/ai00_rwkv_server), change default vocabs(tokenizer) by [this one](https://huggingface.co/xiaol/RWKV-v5-12B-one-state-chat-16k/blob/main/rwkv_vocab_v20230424.txt)
29
+
30
+ # Important Notes
31
+ After overfitting certain instructions and weakening others, it is necessary to use completion or simulate dialogues.
32
+
33
+ - **completion prompt** = 'User: make this content longer:\nxxxxxx\n\nAssistant: ok, longer content is'
34
+
35
+ # Data format
36
+ `<s>User:xxxx\n\n</s>Assistant:xxx\n\n</s>User:xxxx\n\n</s>Assistant:xxx\n\n</s>`
37
+
38
+ If you desire optimal performance to run this model,utilize this format and these [vocabs](https://huggingface.co/xiaol/RWKV-v5-12B-one-state-chat-16k/blob/main/rwkv_vocab_v20230424_train.txt)