|
--- |
|
language: |
|
- el |
|
- en |
|
license: apache-2.0 |
|
pipeline_tag: text-generation |
|
tags: |
|
- finetuned |
|
inference: true |
|
base_model: |
|
- ilsp/Meltemi-7B-Instruct-v1.5 |
|
--- |
|
|
|
# Meltemi 7B Instruct v1.5 gguf |
|
|
|
This is [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5), the first Greek Large Language Model (LLM) published in the `gguf`, [llama.cpp](https://github.com/ggerganov/llama.cpp)-compatible format. |
|
|
|
# Model Information |
|
|
|
- Vocabulary extension of the Mistral 7b tokenizer with Greek tokens for lower costs and faster inference (**1.52** vs. 6.80 tokens/word for Greek) |
|
- 8192 context length |
|
|
|
For more details, please refer to the original model card [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5) |
|
|
|
# Instruction format |
|
|
|
You can do whatever you can with a standard [llama.cpp](https://github.com/ggerganov/llama.cpp) model |
|
|
|
## Basic Usage |
|
|
|
```shell |
|
llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf -p "Ποιό είναι το νόημα της ζωής;" -n 128 |
|
``` |
|
|
|
## Conversation Mode |
|
|
|
```shell |
|
llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --conv |
|
``` |
|
|
|
## Web Server |
|
|
|
```shell |
|
llama-server -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --port 8080 |
|
``` |
|
|
|
|
|
|