qwp4w3hyb's picture
Update README.md
860a91e verified
|
raw
history blame
573 Bytes
metadata
license: gemma
language:
  - en
pipeline_tag: text-generation
tags:
  - google
  - gemma
  - gguf
  - imatrix
base_model: google/gemma-2-9b-it

Quant Infos

  • f32 gguf is from the official kaggle repo
  • imatrix quants are running and will be uploaded one-by-one
  • you will need the gemma2 llama.cpp PR applied to your llama.cpp
  • current quants are based on the f32 gguf provided by google directly, I will reconvert from the huggingface repo once the dust settles to get better gguf metadata

Original Model Card

TODO