--- language: - el - en license: apache-2.0 pipeline_tag: text-generation tags: - finetuned inference: true base_model: - ilsp/Meltemi-7B-Instruct-v1.5 --- # Meltemi 7B Instruct v1.5 gguf This is [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5), the first Greek Large Language Model (LLM) published in the `gguf`, [llama.cpp](https://github.com/ggerganov/llama.cpp)-compatible format. # Model Information - Vocabulary extension of the Mistral 7b tokenizer with Greek tokens for lower costs and faster inference (**1.52** vs. 6.80 tokens/word for Greek) - 8192 context length For more details, please refer to the original model card [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5) # Instruction format You can do whatever you can with a standard [llama.cpp](https://github.com/ggerganov/llama.cpp) model ## Basic Usage ```shell llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf -p "Ποιό είναι το νόημα της ζωής;" -n 128 ``` ## Conversation Mode ```shell llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --conv ``` ## Web Server ```shell llama-server -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --port 8080 ```