Quantization conversion error with LlamaCpp
#11
by
martineden
- opened
During the conversion to GGUF format, received this error:
Traceback (most recent call last):
File "/content/llama.cpp/convert.py", line 1555, in
main()
File "/content/llama.cpp/convert.py", line 1498, in main
params = Params.load(model_plus)
File "/content/llama.cpp/convert.py", line 328, in load
params = Params.loadHFTransformerJson(model_plus.model, hf_config_path)
File "/content/llama.cpp/convert.py", line 237, in loadHFTransformerJson
raise NotImplementedError(f'Unknown rope scaling type: {typ}')
NotImplementedError: Unknown rope scaling type: longrope
is there a way to work around?
Phi-3-mini-128k-instruct is not supported in llama.cpp.
gugarosa
changed discussion status to
closed