--- license: apache-2.0 --- License: Apache-2.0 license under the condition that the model is not used to compete with OpenAI (since that's what Nexusflow said) This is a llamafile for [Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta#starling-lm-7b-beta) quantized to q5-k-m. Safetensors came directly from their [source](https://huggingface.co/Nexusflow/Starling-LM-7B-beta/tree/main), converted to gguf and then quantized to q5-k-m with [llama.cpp](https://github.com/ggerganov/llama.cpp), and then zipped into a llamafile using [Mozilla's awesome project](https://github.com/Mozilla-Ocho/llamafile). -= Llamafile =- Llamafiles are a standalone executable that run an LLM server locally on a variety of operating systems including FreeBSD, Windows, Windows via WSL, Linux, and Mac. The same file works everywhere, I've tested several of these on FreeBSD, Windows, Windows via WSL, and Linux. You just download the .llamafile, (chmod +x or rename to .exe as needed), run it, open the chat interface in a browser, and interact. Options can be passed in to expose the api etc. See their [docs](https://github.com/Mozilla-Ocho/llamafile) for details. [Mozilla Blog Announcement for Llamafile](https://hacks.mozilla.org/2023/11/introducing-llamafile/) - Windows note: If it's over 4gb and you want to use it on Windows, you'll have to run it from WSL. - WSL note: If you get the error about APE, and the recommended command `sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop'` doesn't work, the WSLInterop file might be named something else. I had success with `sudo sh -c 'echo -1 > /proc/sys/fs/binfmt_misc/WSLInterop-late'` If that fails too, just navigate to `/proc/sys/fs/binfmt_msc` and see what files look like `WSLInterop` and echo a -1 to whatever they're called by changing that part of the recommended command. - FreeBSD note: Yes, it actually works on a fresh install of FreeBSD.