morriszms commited on
Commit
710fbae
1 Parent(s): 2b08611

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,16 +1,17 @@
1
  ---
2
- language:
3
- - en
4
- library_name: transformers
5
  license: gemma
 
 
 
 
 
 
 
6
  tags:
7
- - unsloth
8
- - transformers
9
- - gemma2
10
- - gemma
11
  - TensorBlock
12
  - GGUF
13
- base_model: unsloth/gemma-2-2b-it
14
  ---
15
 
16
  <div style="width: auto; margin-left: auto; margin-right: auto">
@@ -24,13 +25,12 @@ base_model: unsloth/gemma-2-2b-it
24
  </div>
25
  </div>
26
 
27
- ## unsloth/gemma-2-2b-it - GGUF
28
 
29
- This repo contains GGUF format model files for [unsloth/gemma-2-2b-it](https://huggingface.co/unsloth/gemma-2-2b-it).
30
 
31
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
32
 
33
-
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
@@ -39,7 +39,6 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
39
 
40
  ## Prompt template
41
 
42
-
43
  ```
44
  <bos><start_of_turn>user
45
  {prompt}<end_of_turn>
@@ -50,18 +49,18 @@ The files were quantized using machines provided by [TensorBlock](https://tensor
50
 
51
  | Filename | Quant type | File Size | Description |
52
  | -------- | ---------- | --------- | ----------- |
53
- | [gemma-2-2b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q2_K.gguf) | Q2_K | 1.145 GB | smallest, significant quality loss - not recommended for most purposes |
54
- | [gemma-2-2b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_S.gguf) | Q3_K_S | 1.267 GB | very small, high quality loss |
55
- | [gemma-2-2b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_M.gguf) | Q3_K_M | 1.361 GB | very small, high quality loss |
56
- | [gemma-2-2b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_L.gguf) | Q3_K_L | 1.444 GB | small, substantial quality loss |
57
- | [gemma-2-2b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_0.gguf) | Q4_0 | 1.518 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
58
- | [gemma-2-2b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_S.gguf) | Q4_K_S | 1.526 GB | small, greater quality loss |
59
- | [gemma-2-2b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_M.gguf) | Q4_K_M | 1.591 GB | medium, balanced quality - recommended |
60
- | [gemma-2-2b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_0.gguf) | Q5_0 | 1.753 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
61
- | [gemma-2-2b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_S.gguf) | Q5_K_S | 1.753 GB | large, low quality loss - recommended |
62
- | [gemma-2-2b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_M.gguf) | Q5_K_M | 1.791 GB | large, very low quality loss - recommended |
63
- | [gemma-2-2b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q6_K.gguf) | Q6_K | 2.004 GB | very large, extremely low quality loss |
64
- | [gemma-2-2b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q8_0.gguf) | Q8_0 | 2.593 GB | very large, extremely low quality loss - not recommended |
65
 
66
 
67
  ## Downloading instruction
 
1
  ---
 
 
 
2
  license: gemma
3
+ library_name: transformers
4
+ pipeline_tag: text-generation
5
+ extra_gated_heading: Access Gemma on Hugging Face
6
+ extra_gated_prompt: To access Gemma on Hugging Face, you’re required to review and
7
+ agree to Google’s usage license. To do this, please ensure you’re logged in to Hugging
8
+ Face and click below. Requests are processed immediately.
9
+ extra_gated_button_content: Acknowledge license
10
  tags:
11
+ - conversational
 
 
 
12
  - TensorBlock
13
  - GGUF
14
+ base_model: google/gemma-2-2b-it
15
  ---
16
 
17
  <div style="width: auto; margin-left: auto; margin-right: auto">
 
25
  </div>
26
  </div>
27
 
28
+ ## google/gemma-2-2b-it - GGUF
29
 
30
+ This repo contains GGUF format model files for [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it).
31
 
32
  The files were quantized using machines provided by [TensorBlock](https://tensorblock.co/), and they are compatible with llama.cpp as of [commit b4011](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d).
33
 
 
34
  <div style="text-align: left; margin: 20px 0;">
35
  <a href="https://tensorblock.co/waitlist/client" style="display: inline-block; padding: 10px 20px; background-color: #007bff; color: white; text-decoration: none; border-radius: 5px; font-weight: bold;">
36
  Run them on the TensorBlock client using your local machine ↗
 
39
 
40
  ## Prompt template
41
 
 
42
  ```
43
  <bos><start_of_turn>user
44
  {prompt}<end_of_turn>
 
49
 
50
  | Filename | Quant type | File Size | Description |
51
  | -------- | ---------- | --------- | ----------- |
52
+ | [gemma-2-2b-it-Q2_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q2_K.gguf) | Q2_K | 1.230 GB | smallest, significant quality loss - not recommended for most purposes |
53
+ | [gemma-2-2b-it-Q3_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_S.gguf) | Q3_K_S | 1.361 GB | very small, high quality loss |
54
+ | [gemma-2-2b-it-Q3_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_M.gguf) | Q3_K_M | 1.462 GB | very small, high quality loss |
55
+ | [gemma-2-2b-it-Q3_K_L.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q3_K_L.gguf) | Q3_K_L | 1.550 GB | small, substantial quality loss |
56
+ | [gemma-2-2b-it-Q4_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_0.gguf) | Q4_0 | 1.630 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
57
+ | [gemma-2-2b-it-Q4_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_S.gguf) | Q4_K_S | 1.639 GB | small, greater quality loss |
58
+ | [gemma-2-2b-it-Q4_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q4_K_M.gguf) | Q4_K_M | 1.709 GB | medium, balanced quality - recommended |
59
+ | [gemma-2-2b-it-Q5_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_0.gguf) | Q5_0 | 1.883 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
60
+ | [gemma-2-2b-it-Q5_K_S.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_S.gguf) | Q5_K_S | 1.883 GB | large, low quality loss - recommended |
61
+ | [gemma-2-2b-it-Q5_K_M.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q5_K_M.gguf) | Q5_K_M | 1.923 GB | large, very low quality loss - recommended |
62
+ | [gemma-2-2b-it-Q6_K.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q6_K.gguf) | Q6_K | 2.151 GB | very large, extremely low quality loss |
63
+ | [gemma-2-2b-it-Q8_0.gguf](https://huggingface.co/tensorblock/gemma-2-2b-it-GGUF/blob/main/gemma-2-2b-it-Q8_0.gguf) | Q8_0 | 2.784 GB | very large, extremely low quality loss - not recommended |
64
 
65
 
66
  ## Downloading instruction
gemma-2-2b-it-Q2_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e84772a111c110902cae9e47a3c8dfa7da4ffe7805a201505d39149250ae33f
3
- size 1229829856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2106abd05a349bcd942c61c183011ef0f41d37ca40eb2d3946fe70c0eb0e51d4
3
+ size 1229829984
gemma-2-2b-it-Q3_K_L.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:510feb77f796f3166322f7fbd56ecb25b478e22b04e8f9eaca370e09210cde8c
3
- size 1550436064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dee119ab3f64b3048a12f48688e423daa14b0f082b8944610cb2a4bac97b21d9
3
+ size 1550436192
gemma-2-2b-it-Q3_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ff458ab88e45028ab57f07600b3304fc916e99e84d8f982346232736c5dd5e4
3
- size 1461667552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10e315a79671d6000d41f2bc38a87286a0155f4b834e3c739d7a6775da40648d
3
+ size 1461667680
gemma-2-2b-it-Q3_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5a25fd22de378196703e8b421db03df23b06531f3e51555340a09890e82bcac
3
- size 1360660192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13b140f5442336fbd22cdeed986404ad13e8328bd296f9f28d4479841c1dba6c
3
+ size 1360660320
gemma-2-2b-it-Q4_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7f64daa5c0f8db69f148b32d2cce037ed3b0012f6773234356342bf7d2e421b8
3
- size 1629509344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4444fb45371ff6515760f3b3f43eb0e928d455d8717e085b63c52257502cf17c
3
+ size 1629509472
gemma-2-2b-it-Q4_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:095ab3fffa40fc601556853149c05a81013275eb33baa4391a124f266ffa802b
3
- size 1708582624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ba0724deef366497c236540b00a189f50432e53326fa32e545ac1e4fabdde3a
3
+ size 1708582752
gemma-2-2b-it-Q4_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b608b04edc196b034777edd6df99cd7dbc1a1de3abb8a8c5dd0dd405323cdbb4
3
- size 1638651616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b9fbe728f523fc020b234ebd08c564c2ac036159a3038679d40771c14cde75b
3
+ size 1638651744
gemma-2-2b-it-Q5_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe9d7625bb96a2a1001bb7da5c3481a96c0e2ace57c1824b3b55c8a4699d30ff
3
- size 1882543840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d95cdc75e692c6a76e0cb73e32825502c91c9773da077da43d9dfb42e5f4c1f
3
+ size 1882543968
gemma-2-2b-it-Q5_K_M.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6441324a0d32bfc8a2a10f27d49c8f7615a1f972ec49ab2aa01dd007981e2111
3
- size 1923278560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99dd96c758984e13ba0adb2749b336e7f13534d6f628ec6ada40e6444bac36f5
3
+ size 1923278688
gemma-2-2b-it-Q5_K_S.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dfe24ec08c75eec1bac8127213a75f6594c6b1dc5b5cb7ddc651e037461cd0c
3
- size 1882543840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8088aed4b2dcddaeb77bf285577931ab79d8ca7653d07462cec125a8f3eefdc
3
+ size 1882543968
gemma-2-2b-it-Q6_K.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cb5606a2b05cebea36ac1c1a80248f612f304c7118880c65529b064f8852aef
3
- size 2151392992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9aaa72603a4dcbd5b68a1ec4a4f05c2515bd04f7e4e6e5c626462c13fbfc639
3
+ size 2151393120
gemma-2-2b-it-Q8_0.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d4470028f5cf2340cda8b4ce1413e52b3ed03ae0938ea5c3c14780c3db9ff3cd
3
- size 2784495328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c23f8e7a88c4a7113f214b672c0b6cb5586f9becbddba6922e172a64528e81c7
3
+ size 2784495456