Supporting GPTQ/AWQ

#2
by fzyzcjy - opened

Hi thank you for this GGUF! I would like to run it in vllm, thus I wonder whether there will be GPTQ/AWQ quantization in the future?

EDIT: Oh I see it seems you only quantize to format GGUF... Then I wonder whether there will be some future plans to also quantize to GPTQ/AWQ

fzyzcjy changed discussion title from Supporting GPTQ to Supporting GPTQ/AWQ

Sign up or log in to comment