TheBloke commited on
Commit
82a762c
1 Parent(s): 4b44675

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -13
README.md CHANGED
@@ -8,8 +8,9 @@ This version has then been quantized to 4-bit using [GPTQ-for-LLaMa](https://git
8
 
9
  ## Other Koala repos
10
 
11
- I have also made these other Koala repose available:
12
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
 
13
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
14
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
15
 
@@ -18,17 +19,24 @@ I have also made these other Koala repose available:
18
  This GPTQ model was quantized using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) with the following command:
19
  ```
20
  python3 llama.py /content/koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save /content/koala-13B-4bit-128g.pt
 
21
  ```
22
 
23
- I created this model using the latest Triton branch of GPTQ-for-LLaMa but I believe it can be run with the CUDA branch also.
24
 
25
- ## How to run with text-generation-webui
26
 
27
- The model files provided will not load as-is with [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
28
 
29
- They require the latest version of the GPTQ code.
30
 
31
- Here are the commands I used to clone GPTQ-for-LLaMa, clone text-generation-webui, and install GPTQ into the UI:
 
 
 
 
 
 
32
  ```
33
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa
34
  git clone https://github.com/oobabooga/text-generation-webui
@@ -36,24 +44,22 @@ mkdir -p text-generation-webui/repositories
36
  ln -s GPTQ-for-LLaMa text-generation-webui/repositories/GPTQ-for-LLaMa
37
  ```
38
 
39
- Then install this model into `text-generation-webui/models` and run text-generation-webui as follows:
40
  ```
41
  cd text-generation-webui
42
- python server.py --model koala-13B-GPTQ-4bit-128g --wbits 4 --groupsize 128 --model_type Llama
43
  ```
44
 
45
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
46
 
47
- If you cannot use the Triton branch for any reason, I believe it should also work to use the CUDA branch instead:
48
  ```
49
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
 
 
50
  ```
51
  Then link that into `text-generation-webui/repositories` as described above.
52
 
53
- ## Coming soon
54
-
55
- Tomorrow I will upload a `safetensors` file as well.
56
-
57
  ## How the Koala delta weights were merged
58
 
59
  The Koala delta weights were originally merged using the following commands, producing [koala-13B-HF](https://huggingface.co/TheBloke/koala-13B-HF):
 
8
 
9
  ## Other Koala repos
10
 
11
+ I have also made these other Koala models available:
12
  * [Unquantized 13B model in HF format](https://huggingface.co/TheBloke/koala-13B-HF)
13
+ * [GPTQ quantized 4bit 7B model in `pt` and `safetensors` formats](https://huggingface.co/TheBloke/koala-7B-4bit-128g)
14
  * [Unquantized 7B model in HF format](https://huggingface.co/TheBloke/koala-7B-HF)
15
  * [Unquantized 7B model in GGML format for llama.cpp](https://huggingface.co/TheBloke/koala-7b-ggml-unquantized)
16
 
 
19
  This GPTQ model was quantized using [GPTQ-for-LLaMa](https://github.com/qwopqwop200/GPTQ-for-LLaMa) with the following command:
20
  ```
21
  python3 llama.py /content/koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save /content/koala-13B-4bit-128g.pt
22
+ python3 llama.py /content/koala-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/koala-13B-4bit-128g.safetensors
23
  ```
24
 
25
+ I created this model using the latest Triton branch of GPTQ-for-LLaMa but it can be loaded with the CUDA branch also.
26
 
27
+ ## Provided files
28
 
29
+ I have provided both a `pt` and `safetensors` file. Either should work.
30
 
31
+ If both are present in the model directory for text-generation-webui I am not sure which it chooses, so you may want to place only one in the models folder.
32
 
33
+ ## How to run with `text-generation-webui`
34
+
35
+ GPTQ model files provided will not load as-is with [oobaboogas text-generation-webui](https://github.com/oobabooga/text-generation-webui).
36
+
37
+ These model files require the latest version of the GPTQ code.
38
+
39
+ Here are the commands I used to clone the Triton branch of GPTQ-for-LLaMa, clone text-generation-webui, and install GPTQ into the UI:
40
  ```
41
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa
42
  git clone https://github.com/oobabooga/text-generation-webui
 
44
  ln -s GPTQ-for-LLaMa text-generation-webui/repositories/GPTQ-for-LLaMa
45
  ```
46
 
47
+ Then install this model into `text-generation-webui/models` and launch the UI as follows:
48
  ```
49
  cd text-generation-webui
50
+ python server.py --model koala-13B-GPTQ-4bit-128g --wbits 4 --groupsize 128 --model_type Llama # add any other command line args you want
51
  ```
52
 
53
  The above commands assume you have installed all dependencies for GPTQ-for-LLaMa and text-generation-webui. Please see their respective repositories for further information.
54
 
55
+ If you cannot use the Triton branch of GPTQ for any reason, you can alternatively use the CUDA branch instead:
56
  ```
57
  git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa -b cuda
58
+ cd GPTQ-for-LLaMa
59
+ python setup_cuda.py install
60
  ```
61
  Then link that into `text-generation-webui/repositories` as described above.
62
 
 
 
 
 
63
  ## How the Koala delta weights were merged
64
 
65
  The Koala delta weights were originally merged using the following commands, producing [koala-13B-HF](https://huggingface.co/TheBloke/koala-13B-HF):