Spaces:
Sleeping
Sleeping
--- | |
title: Introduction | |
--- | |
**Open Interpreter** works with both hosted and local language models. | |
Hosted models are faster and more capable, but require payment. Local models are private and free, but are often less capable. | |
For this reason, we recommend starting with a **hosted** model, then switching to a local model once you've explored Open Interpreter's capabilities. | |
<CardGroup> | |
<Card | |
title="Hosted setup" | |
icon="cloud" | |
href="/language-models/hosted-models" | |
> | |
Connect to a hosted language model like GPT-4 **(recommended)** | |
</Card> | |
<Card | |
title="Local setup" | |
icon="microchip" | |
href="/language-models/local-models" | |
> | |
Setup a local language model like Mistral | |
</Card> | |
</CardGroup> | |
<br/> | |
<br/> | |
<span class="opacity-50">Thank you to the incredible [LiteLLM](https://litellm.ai/) team for their efforts in connecting Open Interpreter to hosted providers.</span> |