Text Generation
Transformers
PyTorch
English
llama
text-generation-inference
Inference Endpoints

Adding `safetensors` variant of this model

#1

This is an automated PR created with https://huggingface.co/spaces/safetensors/convert

This new file is equivalent to pytorch_model.bin but safe in the sense that
no arbitrary code can be put into it.

These files also happen to load much faster than their pytorch counterpart:
https://colab.research.google.com/github/huggingface/notebooks/blob/main/safetensors_doc/en/speed.ipynb

The widgets on your model page will run using this model even if this is not merged
making sure the file actually works.

If you find any issues: please report here: https://huggingface.co/spaces/safetensors/convert/discussions

Feel free to ignore this PR.

Hi authors! Can you please merge this PR so that this model can be easily used with AWQ and other quantization schemes?

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment