quantize my fixed model

#374
by Guilherme34 - opened

Guilherme34/Reasoning-2.6-1b

now its fixed, you can quantize now

Queued. Good luck :)

mradermacher changed discussion status to closed

Alas:

INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only
INFO:hf-to-gguf:Exporting model...
INFO:hf-to-gguf:rope_freqs.weight,           torch.float32 --> F32, shape = {32}
INFO:hf-to-gguf:gguf: loading model part 'model.safetensors'
INFO:hf-to-gguf:token_embd.weight,           torch.float16 --> F16, shape = {2048, 128256}
INFO:hf-to-gguf:blk.0.attn_norm.weight,      torch.float16 --> F32, shape = {2048}
Traceback (most recent call last):
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 4430, in <module>
    main()
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 4424, in main
    model_instance.write()
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 433, in write
    self.prepare_tensors()
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 1653, in prepare_tensors
    super().prepare_tensors()
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 297, in prepare_tensors
    for new_name, data_torch in (self.modify_tensors(data_torch, name, bid)):
                                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 1621, in modify_tensors
    return [(self.map_tensor_name(name), data_torch)]
             ^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "/root/cvs/llama.cpp/convert_hf_to_gguf.py", line 213, in map_tensor_name
    raise ValueError(f"Can not map tensor {name!r}")
ValueError: Can not map tensor 'model.layers.0.mlp.down_proj.SCB'

Sign up or log in to comment