是否支持通过llama.ccp 转换为guff格式,我希望量化后放在ollama中使用
#11
by
ipa2800
- opened
python llm/llama.cpp/convert-hf-to-gguf.py bce-embedding-base_v1 --outtype f16 --outfile bce-embedding-base_v1_converted/bce-embedding-base_v1.bin
INFO:hf-to-gguf:Loading model: bce-embedding-base_v1
Traceback (most recent call last):
File "/llm/llama.cpp/convert-hf-to-gguf.py", line 2562, in <module>
main()
File "/llm/llama.cpp/convert-hf-to-gguf.py", line 2540, in main
model_class = Model.from_model_architecture(hparams["architectures"][0])
File "/llm/llama.cpp/convert-hf-to-gguf.py", line 382, in from_model_architecture
raise NotImplementedError(f'Architecture {arch!r} not supported!') from None
NotImplementedError: Architecture 'XLMRobertaModel' not supported!
the same, I need gguf format