GGUF Please
#1
by
HR1777
- opened
llama.cpp
support for Mamba is coming soon, see https://github.com/ggerganov/llama.cpp/pull/5328
Converting requires adding at least "architectures": ["MambaForCausalLM"],
to config.json
, though.
Got merged
@jondurbin
Please add the missing "architectures": ["MambaForCausalLM"],
line to the config.json
, so that it can be quantized with llama.cpp without any further manipulation.