Updating model files
Browse files
README.md
CHANGED
@@ -10,6 +10,17 @@ tags:
|
|
10 |
- gpt4
|
11 |
inference: false
|
12 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
|
14 |
This is a 4bit 128g GPTQ of [chansung's gpt4-alpaca-lora-13b](https://huggingface.co/chansung/gpt4-alpaca-lora-13b).
|
15 |
|
@@ -31,7 +42,7 @@ Open the text-generation-webui UI as normal.
|
|
31 |
|
32 |
Command to create was:
|
33 |
```
|
34 |
-
CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
|
35 |
```
|
36 |
|
37 |
Command to clone the latest Triton GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
|
@@ -47,6 +58,17 @@ git clone https://github.com/qwopqwop200/GPTQ-for-LLaMa
|
|
47 |
|
48 |
There is also a `no-act-order.safetensors` file which will work with oobabooga's fork of GPTQ-for-LLaMa; it does not require the latest GPTQ code.
|
49 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
# Original model card is below
|
51 |
|
52 |
This repository comes with LoRA checkpoint to make LLaMA into a chatbot like language model. The checkpoint is the output of instruction following fine-tuning process with the following settings on 8xA100(40G) DGX system.
|
|
|
10 |
- gpt4
|
11 |
inference: false
|
12 |
---
|
13 |
+
<div style="width: 100%;">
|
14 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
15 |
+
</div>
|
16 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
17 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
18 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
19 |
+
</div>
|
20 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
21 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
22 |
+
</div>
|
23 |
+
</div>
|
24 |
|
25 |
This is a 4bit 128g GPTQ of [chansung's gpt4-alpaca-lora-13b](https://huggingface.co/chansung/gpt4-alpaca-lora-13b).
|
26 |
|
|
|
42 |
|
43 |
Command to create was:
|
44 |
```
|
45 |
+
CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
|
46 |
```
|
47 |
|
48 |
Command to clone the latest Triton GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
|
|
|
58 |
|
59 |
There is also a `no-act-order.safetensors` file which will work with oobabooga's fork of GPTQ-for-LLaMa; it does not require the latest GPTQ code.
|
60 |
|
61 |
+
## Want to support my work?
|
62 |
+
|
63 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
64 |
+
|
65 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
66 |
+
|
67 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
68 |
+
|
69 |
+
* Patreon: coming soon! (just awaiting approval)
|
70 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
71 |
+
* Discord: https://discord.gg/UBgz4VXf
|
72 |
# Original model card is below
|
73 |
|
74 |
This repository comes with LoRA checkpoint to make LLaMA into a chatbot like language model. The checkpoint is the output of instruction following fine-tuning process with the following settings on 8xA100(40G) DGX system.
|