adamo1139 commited on
Commit
247400e
1 Parent(s): f8be23a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +34 -0
README.md CHANGED
@@ -3,3 +3,37 @@ license: other
3
  license_name: yi-license
4
  license_link: LICENSE
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  license_name: yi-license
4
  license_link: LICENSE
5
  ---
6
+
7
+ ## Model description
8
+
9
+ Yi-34B model fine-tuned on AEZAKMI v1 dataset that is derived from airoboros 2.2.1 and airoboros 2.2. Finetuned with axolotl, using qlora and nf4 double quant, 1 epoch, batch size 1, lr 0.00007, lr scheduler constant. Training took around 33 hours on single local RTX 3090 Ti.
10
+ I had power target set to 320W for the GPU, and while I didn't measure power at the wall, it was probably something around 500W. Given the average electricity price in my region, this training run cost me around $3. This was my first attempt at training Yi-34B with this dataset.
11
+ Main feature of this model is that it's output should be free of refusals and it feels somehow more natural than airoboros. Prompt format is standard chatml. Don't expect it to be good at math, riddles or be crazy smart. My end goal with AEZAKMI is to create a cozy free chatbot.
12
+
13
+ ## Prompt Format
14
+
15
+ I recommend using ChatML format, as this was used during fine-tune
16
+ Here's a prompt format you should use, you can set a different system message, model seems to respect that fine, so it wasn't overfitted.
17
+
18
+ ```
19
+ <|im_start|>system
20
+ A chat.<|im_end|>
21
+ <|im_start|>user
22
+ {prompt}<|im_end|>
23
+ <|im_start|>assistant
24
+ ```
25
+
26
+ ## Intended uses & limitations
27
+
28
+ Use is limited by Yi license
29
+
30
+ ## Known Issues
31
+
32
+ I recommend to set repetition penalty to something around 1.05 to avoid repetition. So far I had good experience running this model with temperature 1.2.
33
+ Multi-turn conversations could be a bit better, if you ask it to re-write something with some fixes it will have a tendency to just repeat the previous response verbatim without any improvements - this is especially noticeable with repp 1.0
34
+ There is still some gptslop left - some responses will have last paragraph with text "Remember that bla bla bla", I will try to get rid of it in the next version of the dataset.
35
+ Stories have ChatGPT like paragraph spacing, I will try to introduce a bit more stories that have long paragraphs in the next dataset version.
36
+
37
+ ## Upcoming
38
+
39
+ I will release adapter files and maybe exllama v2 quant shortly.