File size: 1,692 Bytes
5d23053
 
84b5e7b
 
 
 
 
 
 
 
 
 
 
5d23053
84b5e7b
 
 
8e06e03
 
 
 
abea6aa
84b5e7b
 
 
 
c2365c4
84b5e7b
 
 
 
 
 
 
8e06e03
84b5e7b
8e06e03
 
 
 
 
84b5e7b
8e06e03
84b5e7b
8e06e03
84b5e7b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8e06e03
84b5e7b
 
 
 
 
 
 
 
 
 
 
 
8e06e03
84b5e7b
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
---
license: apache-2.0
datasets:
- yahma/alpaca-cleaned
language:
- de
tags:
- llama
- alpaca
- ggml
- german
- deutsch
- zicklein
---

# ---

---



# Zicklein: A german finetuned instructions following LLaMA


## This is a ggml conversion of [Zicklein](https://github.com/avocardio/zicklein) 7B.

## Zicklein itself is a LLaMA finetuned model with a cleaned and german translated [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) [dataset](https://github.com/LEL-A/GerAlpacaDataCleaned).

Currently I have only converted it into **new k-quant method Q5_K_M**. I will gladly make more versions on request.

Other possible quantizations include: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q5_K_M, q6_K

A f-16 version could be found here: [nikuya3/alpaca-lora-7b-german-base-51k-ggml](https://huggingface.co/nikuya3/alpaca-lora-7b-german-base-51k-ggml)

Compatible with **llama.cpp**, but also with:

- **text-generation-webui**
- **KoboldCpp**
- **ParisNeo/GPT4All-UI**
- **llama-cpp-python**
- **ctransformers**

---

## Prompt format

Since this model is based on alpaca dataset, the right prompt formatting should look like this:

```
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.

### Instruction:
{instruction}

### Input:
{input}

### Response:
```


Or **without** addiotional **input**:


```

Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{instruction}

### Response:
```

---

### That's it!


If you have any further questions, feel free to contact me or start a discussion