File size: 2,410 Bytes
f60a82c
 
 
 
 
 
 
 
 
 
 
 
 
4385359
f60a82c
4385359
f60a82c
4385359
 
 
f60a82c
4385359
f60a82c
4385359
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f60a82c
4385359
f60a82c
4385359
f60a82c
4385359
 
 
f60a82c
 
 
 
 
4385359
f60a82c
 
 
 
 
4385359
f60a82c
 
 
 
 
4385359
 
 
 
f60a82c
4385359
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
---
language:
- el
- en
license: apache-2.0
pipeline_tag: text-generation
tags:
- finetuned
inference: true
base_model:
- ilsp/Meltemi-7B-Instruct-v1.5
---

# Meltemi llamafile & gguf

This repo contains `llamafile` and `gguf` file format models for [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5), the first Greek Large Language Model (LLM)

lamafile is a file format introduced by Mozilla Ocho on Nov 20th 2023, 
and it collapses the complexity of an LLM into a single executable file.
This gives you the easiest, fastest way to use Meltemi on Linux, MacOS, Windows, FreeBSD, OpenBSD, and NetBSD systems you control on both AMD64 and ARM64.

It's as simple as this

```shell
wget https://huggingface.co/Florents-Tselai/Meltemi-llamafile/resolve/main/Meltemi-7B-Instruct-v1.5-Q8_0.llamafile
chmod +x Meltemi-7B-Instruct-v1.5-Q8_0.llamafile
```

```shell
./Meltemi-7B-Instruct-v1.5-Q8_0.llamafile
```

This will open a tab with a chatbot and completion interface in your browser. 
For additional help on how it may be used, pass the `--help` flag. 
The server also has an OpenAI API-compatible completions endpoint.

An advanced CLI mode is provided that's useful for shell scripting. 
You can use it by passing the `--cli` flag. For additional help on how it may be used, pass the --help flag.

```shell
./Meltemi-7B-Instruct-v1.5-Q8_0.llamafile -p 'Ποιό είναι το νόημα της ζωής;'
```

To see all available options

```shell
./Meltemi-7B-Instruct-v1.5-Q8_0.llamafile --help
```

## gguf

`gguf` file formats are also available if you're working with llama.cpp [llama.cpp](https://github.com/ggerganov/llama.cpp)

llama.cpp offers quite a lot of options, thus refer to its documentation. 

### Basic Usage

```shell
llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf -p "Ποιό είναι το νόημα της ζωής;" -n 128
```

### Conversation Mode

```shell
llama-cli -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --conv 
```

### Web Server

```shell
llama-server -m ./Meltemi-7B-Instruct-v1.5-F16.gguf --port 8080
```

# Model Information

- Vocabulary extension of the Mistral 7b tokenizer with Greek tokens for lower costs and faster inference (**1.52** vs. 6.80 tokens/word for Greek)
- 8192 context length

For more details, please refer to the original model card [Meltemi 7B Instract v1.5](https://huggingface.co/ilsp/Meltemi-7B-Instruct-v1.5)