---
license: apache-2.0
tags:
- mixtral
- llamafile
- llm
- moe
---
# Mixtral 8X7B Instruct v0.1 - Llamafile 🦙
## Overview
This model card describes the `mixtral-8x7b-instruct-v0.1.Q3_K_M.llamafile`, a single-file executable version of the Mixtral 8X7B Instruct v0.1 model.
It is built upon the original work by TheBloke and Mistral AI, repackaged for ease of use as a standalone application.
See [here](https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF)
Like many of you, i am GPU poor. The goal behind this approach was to have easy access to a good opensourced model with limited GPU resources, like a Macbook Pro M1 32GB.
It's not the full model, but it's the most feasible given the resource constraints - see [here](https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF#provided-files) for notes on performance
## Usage
Because the model is converted to `llamafile`, it can be executed on any OS with no additional installations required.Read more about llamafile [here](https://github.com/Mozilla-Ocho/llamafile).
To use this model, ensure you have execution permissions set:
```bash
chmod +x mixtral-8x7b-instruct-v0.1.Q3_K_M.llamafile
./mixtral-8x7b-instruct-v0.1.Q3_K_M.llamafile
```
See [here](https://github.com/Mozilla-Ocho/llamafile/blob/6423228b5ddd4862a3ab3d275a168692dadf4cdc/llama.cpp/server/README.md) for local API server details.
## Credits and Acknowledgements
This executable is a derivative of TheBloke's original Mixtral model, repurposed for easier deployment. It is licensed under the same terms as TheBloke's model.
## Limitations
As with the original Mixtral model, this executable does not include moderation mechanisms and should be used with consideration for its capabilities and limitations.
## Additional Information
For more detailed instructions and insights, please refer to the original model documentation provided by TheBloke and Mistral AI.