|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- yahma/alpaca-cleaned |
|
language: |
|
- de |
|
tags: |
|
- llama |
|
- alpaca |
|
- ggml |
|
- german |
|
- deutsch |
|
- zicklein |
|
--- |
|
|
|
# --- |
|
|
|
--- |
|
|
|
|
|
|
|
# Zicklein: A german finetuned instructions following LLaMA |
|
|
|
|
|
## This is a ggml conversion of [Zicklein](https://github.com/avocardio/zicklein) 7B. |
|
|
|
## Zicklein itself is a LLaMA finetuned model with a cleaned and german translated [Alpaca](https://github.com/tatsu-lab/stanford_alpaca) [dataset](https://github.com/LEL-A/GerAlpacaDataCleaned). |
|
|
|
Currently I have only converted it into **new k-quant method Q5_K_M**. I will gladly make more versions on request. |
|
|
|
Other possible quantizations include: q2_K, q3_K_S, q3_K_M, q3_K_L, q4_K_S, q4_K_M, q5_K_S, q5_K_M, q6_K |
|
|
|
A f-16 version could be found here: [nikuya3/alpaca-lora-7b-german-base-51k-ggml](https://huggingface.co/nikuya3/alpaca-lora-7b-german-base-51k-ggml) |
|
|
|
Compatible with **llama.cpp**, but also with: |
|
|
|
- **text-generation-webui** |
|
- **KoboldCpp** |
|
- **ParisNeo/GPT4All-UI** |
|
- **llama-cpp-python** |
|
- **ctransformers** |
|
|
|
--- |
|
|
|
## Prompt format |
|
|
|
Since this model is based on alpaca dataset, the right prompt formatting should look like this: |
|
|
|
``` |
|
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. |
|
|
|
### Instruction: |
|
{instruction} |
|
|
|
### Input: |
|
{input} |
|
|
|
### Response: |
|
``` |
|
|
|
|
|
Or **without** addiotional **input**: |
|
|
|
|
|
``` |
|
|
|
Below is an instruction that describes a task. Write a response that appropriately completes the request. |
|
|
|
### Instruction: |
|
{instruction} |
|
|
|
### Response: |
|
``` |
|
|
|
--- |
|
|
|
### That's it! |
|
|
|
|
|
If you have any further questions, feel free to contact me or start a discussion |