Update README.md
Browse files
README.md
CHANGED
@@ -92,6 +92,7 @@ Refer to the Provided Files table below to see what files use which methods, and
|
|
92 |
<!-- compatibility_ggml end -->
|
93 |
|
94 |
## Provided files
|
|
|
95 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
96 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
97 |
| [stablebeluga2.ggmlv3.q2_K.bin](https://huggingface.co/TheBloke/StableBeluga2-GGML/blob/main/stablebeluga2.ggmlv3.q2_K.bin) | q2_K | 2 | 28.59 GB| 31.09 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. |
|
|
|
92 |
<!-- compatibility_ggml end -->
|
93 |
|
94 |
## Provided files
|
95 |
+
|
96 |
| Name | Quant method | Bits | Size | Max RAM required | Use case |
|
97 |
| ---- | ---- | ---- | ---- | ---- | ----- |
|
98 |
| [stablebeluga2.ggmlv3.q2_K.bin](https://huggingface.co/TheBloke/StableBeluga2-GGML/blob/main/stablebeluga2.ggmlv3.q2_K.bin) | q2_K | 2 | 28.59 GB| 31.09 GB | New k-quant method. Uses GGML_TYPE_Q4_K for the attention.vw and feed_forward.w2 tensors, GGML_TYPE_Q2_K for the other tensors. |
|