qwp4w3hyb commited on
Commit
860a91e
1 Parent(s): 4eb999e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -16,7 +16,7 @@ base_model: google/gemma-2-9b-it
16
  - f32 gguf is from the official kaggle repo
17
  - imatrix quants are running and will be uploaded one-by-one
18
  - you will need the gemma2 llama.cpp [PR](https://github.com/ggerganov/llama.cpp/pull/8156) applied to your llama.cpp
19
- - current quants are based on the f32 gguf provided by google directly, I will reconvert from transformers once the dust settles to get better gguf metadata
20
 
21
  # Original Model Card
22
 
 
16
  - f32 gguf is from the official kaggle repo
17
  - imatrix quants are running and will be uploaded one-by-one
18
  - you will need the gemma2 llama.cpp [PR](https://github.com/ggerganov/llama.cpp/pull/8156) applied to your llama.cpp
19
+ - current quants are based on the f32 gguf provided by google directly, I will reconvert from the huggingface repo once the dust settles to get better gguf metadata
20
 
21
  # Original Model Card
22