TheBloke commited on
Commit
6373fe6
1 Parent(s): 0adce10

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -16
README.md CHANGED
@@ -14,27 +14,32 @@ I have also made these other Koala models available:
14
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
15
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
16
 
17
- ## Quantization method
18
-
19
- This GPTQ model was quantized using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) with the following commands:
20
- ```
21
- python3 llama.py /content/koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save /content/koala-13B-4bit-128g.pt
22
- python3 llama.py /content/koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/koala-13B-4bit-128g.safetensors
23
- ```
24
-
25
- I used the latest Triton branch of `GPTQ-for-LLaMa` but they can also be loaded with the CUDA branch.
26
-
27
  ## Provided files
28
 
29
- I have provided both a `pt` and `safetensors` file. Either should work.
30
 
31
- If both are present in the model directory for text-generation-webui I am not sure which it chooses, so you may want to place only one in the models folder.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
- ## How to run with `text-generation-webui`
34
 
35
- GPTQ model files provided will not load as-is with [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
36
 
37
- These model files require the latest version of the GPTQ code.
38
 
39
  Here are the commands I used to clone the Triton branch of GPTQ-for-LLaMa, clone text-generation-webui, and install GPTQ into the UI:
40
  ```
@@ -52,7 +57,7 @@ python server.py --model koala-13B-GPTQ-4bit-128g --wbits 4 --groupsize 128 --mo
52
 
53
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
54
 
55
- If you cannot use the Triton branch of GPTQ for any reason, you can alternatively use the CUDA branch instead:
56
  ```
57
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
58
  cd GPTQ-for-LLaMa
 
14
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
15
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
16
 
 
 
 
 
 
 
 
 
 
 
17
  ## Provided files
18
 
19
+ Three model files are provided. You don't need all three - choose the one that suits your needs best!
20
 
21
+ Details of the files provided:
22
+ * `koala-13B-4bit-128g.pt`
23
+ * pt format file, created with the latest [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) code.
24
+ * Command to create:
25
+ * `python3 llama.py koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save koala-13B-4bit-128g.pt`
26
+ * `koala-13B-4bit-128g.safetensors`
27
+ * newer `safetensors` format, with improved file security, created with the latest [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) code.
28
+ * Command to create:
29
+ * `python3 llama.py koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors koala-13B-4bit-128g.safetensors`
30
+ * `koala-13B-4bit-128g.no-act-order.ooba.pt`
31
+ * `pt` format file, created with [oobabooga's older CUDA fork of GPTQ-for-LLaMa](https://github.com/oobabooga/GPTQ-for-LLaMa).
32
+ * This file is included primarily for Windows users, as it can be used without needing to compile the latest GPTQ-for-LLaMa code.
33
+ * It should hopefully therefore work with one-click-installers on Windows, which include the older GPTQ-for-LLaMa code.
34
+ * The older GPTQ code does not support all the latest features, so the quality may be fractionally lower.
35
+ * Command to create:
36
+ * `python3 llama.py koala-13B-HF c4 --wbits 4 --true-sequential --groupsize 128 --save koala-13B-4bit-128g.no-act-order.ooba.pt`
37
 
38
+ ## How to run in `text-generation-webui`
39
 
40
+ File `koala-13B-4bit-128g.no-act-order.ooba.pt` can be loaded the same as any other GPTQ file, without requiring any updates to [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
41
 
42
+ The other two model files were created with the latest GPTQ code, and require that the latest GPTQ-for-LLaMa is used inside the UI.
43
 
44
  Here are the commands I used to clone the Triton branch of GPTQ-for-LLaMa, clone text-generation-webui, and install GPTQ into the UI:
45
  ```
 
57
 
58
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
59
 
60
+ If you are on Windows, or cannot use the Triton branch of GPTQ for any other reason, you can instead use the CUDA branch:
61
  ```
62
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
63
  cd GPTQ-for-LLaMa