alexedelsburg
commited on
Commit
•
d9ea9e2
1
Parent(s):
4ae68f2
Update README
Browse files- Fix URL to the model
- List acrastt & Bohan Du
README.md
CHANGED
@@ -9,7 +9,7 @@ pipeline_tag: text-generation
|
|
9 |
---
|
10 |
|
11 |
# Puma 3B - GGUF
|
12 |
-
- Model creator: [Bohan Du](https://huggingface.co/acrastt)
|
13 |
- Original model: [Puma 3B](https://huggingface.co/acrastt/puma-3b)
|
14 |
|
15 |
<!-- description start -->
|
@@ -75,9 +75,9 @@ For other parameters and how to use them, please refer to [the llama.cpp documen
|
|
75 |
|
76 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
77 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
78 |
-
| [puma-3b.q4_1.gguf](https://huggingface.co/
|
79 |
|
80 |
## Thanks
|
81 |
|
82 |
-
- to [Bohan Du](https://huggingface.co/acrastt) for the Puma model
|
83 |
- to [TheBloke](https://huggingface.co/TheBloke) for all the quantized models and this model card template
|
|
|
9 |
---
|
10 |
|
11 |
# Puma 3B - GGUF
|
12 |
+
- Model creator: [Bohan Du / acrastt](https://huggingface.co/acrastt)
|
13 |
- Original model: [Puma 3B](https://huggingface.co/acrastt/puma-3b)
|
14 |
|
15 |
<!-- description start -->
|
|
|
75 |
|
76 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
77 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
78 |
+
| [puma-3b.q4_1.gguf](https://huggingface.co/alexedelsburg/Puma-3b-GGUF/blob/main/puma-3b.q4_1.gguf) | q4_1 | 4 | 2.14 GB| 4.64 GB | Original quant method, 4-bit. Higher accuracy than q4_0 but not as high as q5_0. However has quicker inference than q5 models. |
|
79 |
|
80 |
## Thanks
|
81 |
|
82 |
+
- to [Bohan Du / acrastt](https://huggingface.co/acrastt) for the Puma model
|
83 |
- to [TheBloke](https://huggingface.co/TheBloke) for all the quantized models and this model card template
|