nold commited on
Commit
2d9e3ba
1 Parent(s): 3eca4b5

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ Wiedervereinigung-7b-dpo_Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
37
+ Wiedervereinigung-7b-dpo_Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
38
+ Wiedervereinigung-7b-dpo_Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
39
+ Wiedervereinigung-7b-dpo_Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - DiscoResearch/DiscoLM_German_7b_v1
7
+ - DRXD1000/Phoenix
8
+ - VAGOsolutions/SauerkrautLM-7b-v1-mistral
9
+ - malteos/hermeo-7b
10
+ base_model:
11
+ - DiscoResearch/DiscoLM_German_7b_v1
12
+ - DRXD1000/Phoenix
13
+ - VAGOsolutions/SauerkrautLM-7b-v1-mistral
14
+ - malteos/hermeo-7b
15
+ license: apache-2.0
16
+ language:
17
+ - de
18
+ - en
19
+ ---
20
+
21
+ # Wiedervereinigung-7b-dpo
22
+
23
+ ![image/png](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b/resolve/main/Wiedervereinigung-7b.png)
24
+
25
+ This is a dpo aligned merge of multiple german models scoring 7.1 on the mt-bench-de average.
26
+ It is a merge of the best german 7B models with 7b parameters as a dare_ties merge.
27
+ Since the original models based on mistral - three of them on the brilliant german LeoLM/leo-mistral-hessianai-7b - they are reunited in this merged model.
28
+ Therefore the name, no nationalist ideas involved. To improve result quality they are dpo-trained with a german translation of intel-orca-dpo
29
+ using our german fork of [LLaMA-Factory](https://github.com/mayflower/LLaMA-Factory-de).
30
+
31
+ ## mt-bench-de
32
+
33
+ Is the merged model good? Well, of course. But it is even better with the help of some dpo tuning.
34
+
35
+ ```json
36
+ {
37
+ "first_turn": 7.3,
38
+ "second_turn": 6.925,
39
+ "categories": {
40
+ "writing": 8.425,
41
+ "roleplay": 8.6,
42
+ "reasoning": 5.4,
43
+ "math": 4.35,
44
+ "coding": 4.3,
45
+ "extraction": 7.975,
46
+ "stem": 8.5,
47
+ "humanities": 9.35
48
+ },
49
+ "average": 7.1125
50
+ }
51
+ ```
52
+
53
+
54
+ Wiedervereinigung-7b itself is a [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing) merge of:
55
+ * [DiscoResearch/DiscoLM_German_7b_v1](https://huggingface.co/DiscoResearch/DiscoLM_German_7b_v1)
56
+ * [DRXD1000/Phoenix](https://huggingface.co/DRXD1000/Phoenix)
57
+ * [VAGOsolutions/SauerkrautLM-7b-v1-mistral](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-v1-mistral)
58
+ * [malteos/hermeo-7b](https://huggingface.co/malteos/hermeo-7b)
59
+
60
+ All the actual heavylifting has been done by the creators of these models.
61
+
62
+ ## 🧩 Configuration
63
+
64
+ ```yaml
65
+ models:
66
+ - model: LeoLM/leo-mistral-hessianai-7b
67
+ # No parameters necessary for base model
68
+ - model: DiscoResearch/DiscoLM_German_7b_v1
69
+ parameters:
70
+ density: 0.6
71
+ weight: 0.25
72
+ - model: DRXD1000/Phoenix
73
+ parameters:
74
+ density: 0.6
75
+ weight: 0.25
76
+ - model: VAGOsolutions/SauerkrautLM-7b-v1-mistral
77
+ parameters:
78
+ density: 0.6
79
+ weight: 0.25
80
+ - model: malteos/hermeo-7b
81
+ parameters:
82
+ density: 0.6
83
+ weight: 0.25
84
+ merge_method: dare_ties
85
+ base_model: LeoLM/leo-mistral-hessianai-7b
86
+ parameters:
87
+ int8_mask: true
88
+ dtype: bfloat16
89
+ ```
90
+
91
+
92
+ ## 💻 Usage
93
+
94
+ ```python
95
+ !pip install -qU transformers accelerate
96
+
97
+ from transformers import AutoTokenizer
98
+ import transformers
99
+ import torch
100
+
101
+ model = "mayflowergmbh/Wiedervereinigung-7b-dpo"
102
+ messages = [{"role": "user", "content": "Was ist ein deutsches large language model?"}]
103
+
104
+ tokenizer = AutoTokenizer.from_pretrained(model)
105
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
106
+ pipeline = transformers.pipeline(
107
+ "text-generation",
108
+ model=model,
109
+ torch_dtype=torch.float16,
110
+ device_map="auto",
111
+ )
112
+
113
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
114
+ print(outputs[0]["generated_text"])
115
+ ```
116
+
117
+ ***
118
+
119
+ Vanilla Quantization by [nold](https://huggingface.co/nold), Model by [mayflowergmbh](https://huggingface.co/mayflowergmbh/Wiedervereinigung-7b-dpo). Created using [llm-quantizer](https://github.com/Nold360/llm-quantizer) Pipeline - 4bc844478df79ecfd72815473b30ae09499e179d
Wiedervereinigung-7b-dpo_Q2_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a47e26c1e8845d1dda325c8d12be2cfeaf3ae2b0cce1e0d7192b06375ac5a6cc
3
+ size 2719241888
Wiedervereinigung-7b-dpo_Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2833dbda84818fd379f4a70bacbd3ce9b935c85b10e9e8f7f48980666bdf25c1
3
+ size 4368438944
Wiedervereinigung-7b-dpo_Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8e0e183b42c50f14544632f1aaa595653465c2235d73c515d1f6e3c58d5166f
3
+ size 5131409056
Wiedervereinigung-7b-dpo_Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfd0a0482a0ffb8f8f2dcf803e67e2c801b5c596c42f0813e63ab0a99f3404d0
3
+ size 7695857312