--- license: gemma language: - en pipeline_tag: text-generation tags: - google - gemma - gguf - imatrix base_model: google/gemma-2-9b-it --- # Quant Infos - f32 gguf is from the official kaggle repo - imatrix quants are running and will be uploaded one-by-one - you will need the gemma2 llama.cpp [PR](https://github.com/ggerganov/llama.cpp/pull/8156) applied to your llama.cpp - current quants are based on the f32 gguf provided by google directly, I will reconvert from the huggingface repo once the dust settles to get better gguf metadata # Original Model Card TODO