Update README.md
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ base_model: google/gemma-2-9b-it
|
|
16 |
- f32 gguf is from the official kaggle repo
|
17 |
- imatrix quants are running and will be uploaded one-by-one
|
18 |
- you will need the gemma2 llama.cpp [PR](https://github.com/ggerganov/llama.cpp/pull/8156) applied to your llama.cpp
|
19 |
-
- current quants are based on the f32 gguf provided by google directly, I will reconvert from transformers once the dust settles to get better gguf metadata
|
20 |
|
21 |
# Original Model Card
|
22 |
|
|
|
16 |
- f32 gguf is from the official kaggle repo
|
17 |
- imatrix quants are running and will be uploaded one-by-one
|
18 |
- you will need the gemma2 llama.cpp [PR](https://github.com/ggerganov/llama.cpp/pull/8156) applied to your llama.cpp
|
19 |
+
- current quants are based on the f32 gguf provided by google directly, I will reconvert from transformers once the dust settles to get better gguf metadata
|
20 |
|
21 |
# Original Model Card
|
22 |
|