Neko-Institute-of-Science
commited on
Commit
•
e86ab76
1
Parent(s):
51d69a2
update notes
Browse files
README.md
CHANGED
@@ -11,16 +11,17 @@ https://github.com/oobabooga/text-generation-webui
|
|
11 |
ATM I'm using 2023.05.04v0 of the dataset and training full context.
|
12 |
|
13 |
# Notes:
|
14 |
-
So
|
15 |
-
|
16 |
-
Also I will be uploading checkpoints almost everyday.
|
17 |
|
18 |
# How to test?
|
19 |
1. Download LLaMA-30B-HF: https://huggingface.co/Neko-Institute-of-Science/LLaMA-30B-HF
|
20 |
2. Replace special_tokens_map.json and tokenizer_config.json using the ones on this repo.
|
21 |
3. Rename LLaMA-30B-HF to vicuna-30b
|
22 |
-
4.
|
23 |
-
5.
|
|
|
24 |
|
25 |
|
26 |
# Want to see it Training?
|
|
|
11 |
ATM I'm using 2023.05.04v0 of the dataset and training full context.
|
12 |
|
13 |
# Notes:
|
14 |
+
So I will only be training 1 epoch, as full context 30b takes so long to train.
|
15 |
+
This 1 epoch will take me 8 days lol but luckily these LoRA feels fully functinal at epoch 1 as shown on my 13b one.
|
16 |
+
Also I will be uploading checkpoints almost everyday. I could train another epoch if there's enough want for it.
|
17 |
|
18 |
# How to test?
|
19 |
1. Download LLaMA-30B-HF: https://huggingface.co/Neko-Institute-of-Science/LLaMA-30B-HF
|
20 |
2. Replace special_tokens_map.json and tokenizer_config.json using the ones on this repo.
|
21 |
3. Rename LLaMA-30B-HF to vicuna-30b
|
22 |
+
4. Download the checkpoint-xxxx you want and put it in the loras folder.
|
23 |
+
5. Load ooba: ```python server.py --listen --model vicuna-30b --load-in-8bit --chat --lora checkpoint-xxxx```
|
24 |
+
6. Instruct mode: Vicuna-v1, ooba will load Vicuna-v0 by defualt
|
25 |
|
26 |
|
27 |
# Want to see it Training?
|