Upload folder using huggingface_hub
Browse files- README.md +23 -24
- gemma-2-2b-it-Q2_K.gguf +2 -2
- gemma-2-2b-it-Q3_K_L.gguf +2 -2
- gemma-2-2b-it-Q3_K_M.gguf +2 -2
- gemma-2-2b-it-Q3_K_S.gguf +2 -2
- gemma-2-2b-it-Q4_0.gguf +2 -2
- gemma-2-2b-it-Q4_K_M.gguf +2 -2
- gemma-2-2b-it-Q4_K_S.gguf +2 -2
- gemma-2-2b-it-Q5_0.gguf +2 -2
- gemma-2-2b-it-Q5_K_M.gguf +2 -2
- gemma-2-2b-it-Q5_K_S.gguf +2 -2
- gemma-2-2b-it-Q6_K.gguf +2 -2
- gemma-2-2b-it-Q8_0.gguf +2 -2
README.md
CHANGED
@@ -1,16 +1,17 @@
|
|
1 |
---
|
2 |
-
language:
|
3 |
-
- en
|
4 |
-
library_name: transformers
|
5 |
license: gemma
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
tags:
|
7 |
-
-
|
8 |
-
- transformers
|
9 |
-
- gemma2
|
10 |
-
- gemma
|
11 |
- TensorBlock
|
12 |
- GGUF
|
13 |
-
base_model:
|
14 |
---
|
15 |
|
16 |
<div style="width: auto; margin-left: auto; margin-right: auto">
|
@@ -24,13 +25,12 @@ base_model: unsloth/gemma-2-2b-it
|
|
24 |
</div>
|
25 |
</div>
|
26 |
|
27 |
-
##
|
28 |
|
29 |
-
This repo contains GGUF format model files for [
|
30 |
|
31 |
The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
|
32 |
|
33 |
-
|
34 |
<div style="text-align: left; margin: 20px 0;">
|
35 |
<a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
|
36 |
Run them on the TensorBlock client using your local machine ↗
|
@@ -39,7 +39,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
|
|
39 |
|
40 |
## Prompt template
|
41 |
|
42 |
-
|
43 |
```
|
44 |
<bos><start_of_turn>user
|
45 |
{prompt}<end_of_turn>
|
@@ -50,18 +49,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
|
|
50 |
|
51 |
| Filename | Quant type | File Size | Description |
|
52 |
| -------- | ---------- | --------- | ----------- |
|
53 |
-
| [gemma-2-2b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q2_K.gguf) | Q2_K | 1.
|
54 |
-
| [gemma-2-2b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_S.gguf) | Q3_K_S | 1.
|
55 |
-
| [gemma-2-2b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_M.gguf) | Q3_K_M | 1.
|
56 |
-
| [gemma-2-2b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_L.gguf) | Q3_K_L | 1.
|
57 |
-
| [gemma-2-2b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_0.gguf) | Q4_0 | 1.
|
58 |
-
| [gemma-2-2b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_S.gguf) | Q4_K_S | 1.
|
59 |
-
| [gemma-2-2b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_M.gguf) | Q4_K_M | 1.
|
60 |
-
| [gemma-2-2b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_0.gguf) | Q5_0 | 1.
|
61 |
-
| [gemma-2-2b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_S.gguf) | Q5_K_S | 1.
|
62 |
-
| [gemma-2-2b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_M.gguf) | Q5_K_M | 1.
|
63 |
-
| [gemma-2-2b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q6_K.gguf) | Q6_K | 2.
|
64 |
-
| [gemma-2-2b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q8_0.gguf) | Q8_0 | 2.
|
65 |
|
66 |
|
67 |
## Downloading instruction
|
|
|
1 |
---
|
|
|
|
|
|
|
2 |
license: gemma
|
3 |
+
library_name: transformers
|
4 |
+
pipeline_tag: text-generation
|
5 |
+
extra_gated_heading: Access Gemma on Hugging Face
|
6 |
+
extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
|
7 |
+
agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
|
8 |
+
Face and click below. Requests are processed immediately.
|
9 |
+
extra_gated_button_content: Acknowledge license
|
10 |
tags:
|
11 |
+
- conversational
|
|
|
|
|
|
|
12 |
- TensorBlock
|
13 |
- GGUF
|
14 |
+
base_model: google/gemma-2-2b-it
|
15 |
---
|
16 |
|
17 |
<div style="width: auto; margin-left: auto; margin-right: auto">
|
|
|
25 |
</div>
|
26 |
</div>
|
27 |
|
28 |
+
## google/gemma-2-2b-it - GGUF
|
29 |
|
30 |
+
This repo contains GGUF format model files for [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it).
|
31 |
|
32 |
The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
|
33 |
|
|
|
34 |
<div style="text-align: left; margin: 20px 0;">
|
35 |
<a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
|
36 |
Run them on the TensorBlock client using your local machine ↗
|
|
|
39 |
|
40 |
## Prompt template
|
41 |
|
|
|
42 |
```
|
43 |
<bos><start_of_turn>user
|
44 |
{prompt}<end_of_turn>
|
|
|
49 |
|
50 |
| Filename | Quant type | File Size | Description |
|
51 |
| -------- | ---------- | --------- | ----------- |
|
52 |
+
| [gemma-2-2b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q2_K.gguf) | Q2_K | 1.230 GB | smallest, significant quality loss - not recommended for most purposes |
|
53 |
+
| [gemma-2-2b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_S.gguf) | Q3_K_S | 1.361 GB | very small, high quality loss |
|
54 |
+
| [gemma-2-2b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_M.gguf) | Q3_K_M | 1.462 GB | very small, high quality loss |
|
55 |
+
| [gemma-2-2b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_L.gguf) | Q3_K_L | 1.550 GB | small, substantial quality loss |
|
56 |
+
| [gemma-2-2b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_0.gguf) | Q4_0 | 1.630 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
|
57 |
+
| [gemma-2-2b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_S.gguf) | Q4_K_S | 1.639 GB | small, greater quality loss |
|
58 |
+
| [gemma-2-2b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_M.gguf) | Q4_K_M | 1.709 GB | medium, balanced quality - recommended |
|
59 |
+
| [gemma-2-2b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_0.gguf) | Q5_0 | 1.883 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
|
60 |
+
| [gemma-2-2b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_S.gguf) | Q5_K_S | 1.883 GB | large, low quality loss - recommended |
|
61 |
+
| [gemma-2-2b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_M.gguf) | Q5_K_M | 1.923 GB | large, very low quality loss - recommended |
|
62 |
+
| [gemma-2-2b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q6_K.gguf) | Q6_K | 2.151 GB | very large, extremely low quality loss |
|
63 |
+
| [gemma-2-2b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q8_0.gguf) | Q8_0 | 2.784 GB | very large, extremely low quality loss - not recommended |
|
64 |
|
65 |
|
66 |
## Downloading instruction
|
gemma-2-2b-it-Q2_K.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2106abd05a349bcd942c61c183011ef0f41d37ca40eb2d3946fe70c0eb0e51d4
|
3 |
+
size 1229829984
|
gemma-2-2b-it-Q3_K_L.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dee119ab3f64b3048a12f48688e423daa14b0f082b8944610cb2a4bac97b21d9
|
3 |
+
size 1550436192
|
gemma-2-2b-it-Q3_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:10e315a79671d6000d41f2bc38a87286a0155f4b834e3c739d7a6775da40648d
|
3 |
+
size 1461667680
|
gemma-2-2b-it-Q3_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:13b140f5442336fbd22cdeed986404ad13e8328bd296f9f28d4479841c1dba6c
|
3 |
+
size 1360660320
|
gemma-2-2b-it-Q4_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4444fb45371ff6515760f3b3f43eb0e928d455d8717e085b63c52257502cf17c
|
3 |
+
size 1629509472
|
gemma-2-2b-it-Q4_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ba0724deef366497c236540b00a189f50432e53326fa32e545ac1e4fabdde3a
|
3 |
+
size 1708582752
|
gemma-2-2b-it-Q4_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b9fbe728f523fc020b234ebd08c564c2ac036159a3038679d40771c14cde75b
|
3 |
+
size 1638651744
|
gemma-2-2b-it-Q5_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d95cdc75e692c6a76e0cb73e32825502c91c9773da077da43d9dfb42e5f4c1f
|
3 |
+
size 1882543968
|
gemma-2-2b-it-Q5_K_M.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:99dd96c758984e13ba0adb2749b336e7f13534d6f628ec6ada40e6444bac36f5
|
3 |
+
size 1923278688
|
gemma-2-2b-it-Q5_K_S.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8088aed4b2dcddaeb77bf285577931ab79d8ca7653d07462cec125a8f3eefdc
|
3 |
+
size 1882543968
|
gemma-2-2b-it-Q6_K.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e9aaa72603a4dcbd5b68a1ec4a4f05c2515bd04f7e4e6e5c626462c13fbfc639
|
3 |
+
size 2151393120
|
gemma-2-2b-it-Q8_0.gguf
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c23f8e7a88c4a7113f214b672c0b6cb5586f9becbddba6922e172a64528e81c7
|
3 |
+
size 2784495456
|