LHC88 commited on
Commit
bc5bbc1
1 Parent(s): 65f7bf3

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +426 -0
README.md ADDED
@@ -0,0 +1,426 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct
3
+ datasets:
4
+ - argilla/distilabel-math-preference-dpo
5
+ inference: false
6
+ language:
7
+ - en
8
+ - de
9
+ - fr
10
+ - it
11
+ - es
12
+ library_name: transformers
13
+ license: apache-2.0
14
+ model_creator: VAGO solutions
15
+ model_name: SauerkrautLM Mixtral 8X7B Instruct
16
+ model_type: mixtral
17
+ pipeline_tag: text-generation
18
+ prompt_template: '[INST] {prompt} [/INST]
19
+
20
+ '
21
+ quantized_by: LHC
22
+ tags:
23
+ - mistral
24
+ - finetune
25
+ - dpo
26
+ - Instruct
27
+ - augmentation
28
+ - german
29
+ - mixtral
30
+ ---
31
+ <!-- markdownlint-disable MD041 -->
32
+
33
+ <!-- header start -->
34
+ <!-- 200823 -->
35
+ <div style="width: auto; margin-left: auto; margin-right: auto">
36
+ <img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
37
+ </div>
38
+ <div style="display: flex; justify-content: space-between; width: 100%;">
39
+ <div style="display: flex; flex-direction: column; align-items: flex-start;">
40
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://discord.gg/theblokeai">Chat & support: TheBloke's Discord server</a></p>
41
+ </div>
42
+ <div style="display: flex; flex-direction: column; align-items: flex-end;">
43
+ <p style="margin-top: 0.5em; margin-bottom: 0em;"><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? TheBloke's Patreon page</a></p>
44
+ </div>
45
+ </div>
46
+ <div style="text-align:center; margin-top: 0em; margin-bottom: 0em"><p style="margin-top: 0.25em; margin-bottom: 0em;">TheBloke's LLM work is generously supported by a grant from <a href="https://a16z.com">andreessen horowitz (a16z)</a></p></div>
47
+ <hr style="margin-top: 1.0em; margin-bottom: 1.0em;">
48
+ <!-- header end -->
49
+
50
+ # SauerkrautLM Mixtral 8X7B Instruct - AWQ
51
+ - Model creator: [VAGO solutions](https://huggingface.co/VAGOsolutions)
52
+ - Original model: [SauerkrautLM Mixtral 8X7B Instruct](https://huggingface.co/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct)
53
+
54
+ <!-- description start -->
55
+ ## Description
56
+
57
+ This repo contains AWQ model files for [VAGO solutions's SauerkrautLM Mixtral 8X7B Instruct](https://huggingface.co/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct).
58
+
59
+
60
+ **MIXTRAL AWQ**
61
+
62
+ This is a Mixtral AWQ model. With a slightly better quantisation than THeBoke's AWQ version.
63
+
64
+ For AutoAWQ inference, please install AutoAWQ 0.1.8 or later.
65
+
66
+ Support via Transformers is also available, but currently requires installing Transformers from Github: `pip3 install git+https://github.com/huggingface/transformers.git`
67
+
68
+ vLLM: version 0.2.6 is confirmed to support Mixtral AWQs.
69
+
70
+ TGI: I tested version 1.3.3 and it loaded the model fine, but I was not able to get any output back. Further testing/debug is required. (Let me know if you get it working!)
71
+
72
+ ### About AWQ
73
+
74
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
75
+
76
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
77
+
78
+ AWQ models are supported by (note that not all of these may support Mixtral models yet - see above):
79
+
80
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
81
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
82
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
83
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
84
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
85
+
86
+ <!-- description end -->
87
+ <!-- repositories-available start -->
88
+ ## Repositories available
89
+
90
+ * [AWQ model(s) for GPU inference.](https://huggingface.co/TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ)
91
+ * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-GPTQ)
92
+ * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-GGUF)
93
+ * [VAGO solutions's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct)
94
+ <!-- repositories-available end -->
95
+
96
+ <!-- prompt-template start -->
97
+ ## Prompt template: Mistral
98
+
99
+ ```
100
+ [INST] {prompt} [/INST]
101
+
102
+ ```
103
+
104
+ <!-- prompt-template end -->
105
+
106
+
107
+ <!-- README_AWQ.md-provided-files start -->
108
+ ## Provided files, and AWQ parameters
109
+
110
+ I currently release 128g GEMM models only. The addition of group_size 32 models, and GEMV kernel models, is being actively considered.
111
+
112
+ Models are released as sharded safetensors files.
113
+
114
+ | Branch | Bits | GS | AWQ Dataset | Seq Len | Size |
115
+ | ------ | ---- | -- | ----------- | ------- | ---- |
116
+ | [main](https://huggingface.co/TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ/tree/main) | 4 | 32 | [German Quad](https://huggingface.co/datasets/deepset/germanquad/viewer/) | 8192 | 24.65 GB
117
+
118
+ <!-- README_AWQ.md-provided-files end -->
119
+
120
+ <!-- README_AWQ.md-text-generation-webui start -->
121
+ ## How to easily download and use this model in [text-generation-webui](https://github.com/oobabooga/text-generation-webui)
122
+
123
+ Please make sure you're using the latest version of [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
124
+
125
+ It is strongly recommended to use the text-generation-webui one-click-installers unless you're sure you know how to make a manual install.
126
+
127
+ 1. Click the **Model tab**.
128
+ 2. Under **Download custom model or LoRA**, enter `TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ`.
129
+ 3. Click **Download**.
130
+ 4. The model will start downloading. Once it's finished it will say "Done".
131
+ 5. In the top left, click the refresh icon next to **Model**.
132
+ 6. In the **Model** dropdown, choose the model you just downloaded: `SauerkrautLM-Mixtral-8x7B-Instruct-AWQ`
133
+ 7. Select **Loader: AutoAWQ**.
134
+ 8. Click Load, and the model will load and is now ready for use.
135
+ 9. If you want any custom settings, set them and then click **Save settings for this model** followed by **Reload the Model** in the top right.
136
+ 10. Once you're ready, click the **Text Generation** tab and enter a prompt to get started!
137
+ <!-- README_AWQ.md-text-generation-webui end -->
138
+
139
+ <!-- README_AWQ.md-use-from-vllm start -->
140
+ ## Multi-user inference server: vLLM
141
+
142
+ Documentation on installing and using vLLM [can be found here](https://vllm.readthedocs.io/en/latest/).
143
+
144
+ - Please ensure you are using vLLM version 0.2 or later.
145
+ - When using vLLM as a server, pass the `--quantization awq` parameter.
146
+
147
+ For example:
148
+
149
+ ```shell
150
+ python3 -m vllm.entrypoints.api_server --model TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ --quantization awq --dtype auto
151
+ ```
152
+
153
+ - When using vLLM from Python code, again set `quantization=awq`.
154
+
155
+ For example:
156
+
157
+ ```python
158
+ from vllm import LLM, SamplingParams
159
+
160
+ prompts = [
161
+ "Tell me about AI",
162
+ "Write a story about llamas",
163
+ "What is 291 - 150?",
164
+ "How much wood would a woodchuck chuck if a woodchuck could chuck wood?",
165
+ ]
166
+ prompt_template=f'''[INST] {prompt} [/INST]
167
+ '''
168
+
169
+ prompts = [prompt_template.format(prompt=prompt) for prompt in prompts]
170
+
171
+ sampling_params = SamplingParams(temperature=0.8, top_p=0.95)
172
+
173
+ llm = LLM(model="TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ", quantization="awq", dtype="auto")
174
+
175
+ outputs = llm.generate(prompts, sampling_params)
176
+
177
+ # Print the outputs.
178
+ for output in outputs:
179
+ prompt = output.prompt
180
+ generated_text = output.outputs[0].text
181
+ print(f"Prompt: {prompt!r}, Generated text: {generated_text!r}")
182
+ ```
183
+ <!-- README_AWQ.md-use-from-vllm start -->
184
+
185
+ <!-- README_AWQ.md-use-from-tgi start -->
186
+ ## Multi-user inference server: Hugging Face Text Generation Inference (TGI)
187
+
188
+ Use TGI version 1.1.0 or later. The official Docker container is: `ghcr.io/huggingface/text-generation-inference:1.1.0`
189
+
190
+ Example Docker parameters:
191
+
192
+ ```shell
193
+ --model-id TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ --port 3000 --quantize awq --max-input-length 3696 --max-total-tokens 4096 --max-batch-prefill-tokens 4096
194
+ ```
195
+
196
+ Example Python code for interfacing with TGI (requires [huggingface-hub](https://github.com/huggingface/huggingface_hub) 0.17.0 or later):
197
+
198
+ ```shell
199
+ pip3 install huggingface-hub
200
+ ```
201
+
202
+ ```python
203
+ from huggingface_hub import InferenceClient
204
+
205
+ endpoint_url = "https://your-endpoint-url-here"
206
+
207
+ prompt = "Tell me about AI"
208
+ prompt_template=f'''[INST] {prompt} [/INST]
209
+ '''
210
+
211
+ client = InferenceClient(endpoint_url)
212
+ response = client.text_generation(prompt,
213
+ max_new_tokens=128,
214
+ do_sample=True,
215
+ temperature=0.7,
216
+ top_p=0.95,
217
+ top_k=40,
218
+ repetition_penalty=1.1)
219
+
220
+ print(f"Model output: ", response)
221
+ ```
222
+ <!-- README_AWQ.md-use-from-tgi end -->
223
+
224
+ <!-- README_AWQ.md-use-from-python start -->
225
+ ## Inference from Python code using Transformers
226
+
227
+ ### Install the necessary packages
228
+
229
+ - Requires: [Transformers](https://huggingface.co/docs/transformers) 4.35.0 or later.
230
+ - Requires: [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) 0.1.6 or later.
231
+
232
+ ```shell
233
+ pip3 install --upgrade "autoawq>=0.1.6" "transformers>=4.35.0"
234
+ ```
235
+
236
+ Note that if you are using PyTorch 2.0.1, the above AutoAWQ command will automatically upgrade you to PyTorch 2.1.0.
237
+
238
+ If you are using CUDA 11.8 and wish to continue using PyTorch 2.0.1, instead run this command:
239
+
240
+ ```shell
241
+ pip3 install https://github.com/casper-hansen/AutoAWQ/releases/download/v0.1.6/autoawq-0.1.6+cu118-cp310-cp310-linux_x86_64.whl
242
+ ```
243
+
244
+ If you have problems installing [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) using the pre-built wheels, install it from source instead:
245
+
246
+ ```shell
247
+ pip3 uninstall -y autoawq
248
+ git clone https://github.com/casper-hansen/AutoAWQ
249
+ cd AutoAWQ
250
+ pip3 install .
251
+ ```
252
+
253
+ ### Transformers example code (requires Transformers 4.35.0 and later)
254
+
255
+ ```python
256
+ from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer
257
+
258
+ model_name_or_path = "TheBloke/SauerkrautLM-Mixtral-8x7B-Instruct-AWQ"
259
+
260
+ tokenizer = AutoTokenizer.from_pretrained(model_name_or_path)
261
+ model = AutoModelForCausalLM.from_pretrained(
262
+ model_name_or_path,
263
+ low_cpu_mem_usage=True,
264
+ device_map="cuda:0"
265
+ )
266
+
267
+ # Using the text streamer to stream output one token at a time
268
+ streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
269
+
270
+ prompt = "Tell me about AI"
271
+ prompt_template=f'''[INST] {prompt} [/INST]
272
+ '''
273
+
274
+ # Convert prompt to tokens
275
+ tokens = tokenizer(
276
+ prompt_template,
277
+ return_tensors='pt'
278
+ ).input_ids.cuda()
279
+
280
+ generation_params = {
281
+ "do_sample": True,
282
+ "temperature": 0.7,
283
+ "top_p": 0.95,
284
+ "top_k": 40,
285
+ "max_new_tokens": 512,
286
+ "repetition_penalty": 1.1
287
+ }
288
+
289
+ # Generate streamed output, visible one token at a time
290
+ generation_output = model.generate(
291
+ tokens,
292
+ streamer=streamer,
293
+ **generation_params
294
+ )
295
+
296
+ # Generation without a streamer, which will include the prompt in the output
297
+ generation_output = model.generate(
298
+ tokens,
299
+ **generation_params
300
+ )
301
+
302
+ # Get the tokens from the output, decode them, print them
303
+ token_output = generation_output[0]
304
+ text_output = tokenizer.decode(token_output)
305
+ print("model.generate output: ", text_output)
306
+
307
+ # Inference is also possible via Transformers' pipeline
308
+ from transformers import pipeline
309
+
310
+ pipe = pipeline(
311
+ "text-generation",
312
+ model=model,
313
+ tokenizer=tokenizer,
314
+ **generation_params
315
+ )
316
+
317
+ pipe_output = pipe(prompt_template)[0]['generated_text']
318
+ print("pipeline output: ", pipe_output)
319
+
320
+ ```
321
+ <!-- README_AWQ.md-use-from-python end -->
322
+
323
+ <!-- README_AWQ.md-compatibility start -->
324
+ ## Compatibility
325
+
326
+ The files provided are tested to work with:
327
+
328
+ - [text-generation-webui](https://github.com/oobabooga/text-generation-webui) using `Loader: AutoAWQ`.
329
+ - [vLLM](https://github.com/vllm-project/vllm) version 0.2.0 and later.
330
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference) version 1.1.0 and later.
331
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later.
332
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) version 0.1.1 and later.
333
+
334
+ <!-- README_AWQ.md-compatibility end -->
335
+
336
+ <!-- footer start -->
337
+ <!-- 200823 -->
338
+
339
+ <!-- footer end -->
340
+
341
+ # Original model card: VAGO solutions's SauerkrautLM Mixtral 8X7B Instruct
342
+
343
+
344
+ ![SauerkrautLM](https://vago-solutions.de/wp-content/uploads/2023/12/Sauerkraut_Instruct_MoE_Instruct.png "SauerkrautLM-Mixtral-8x7B")
345
+ ## VAGO solutions SauerkrautLM-Mixtral-8x7B-Instruct
346
+ Introducing **SauerkrautLM-Mixtral-8x7B-Instruct** – our Sauerkraut version of the powerful Mixtral-8x7B-Instruct!
347
+ Aligned with **DPO**
348
+
349
+ # Table of Contents
350
+ 1. [Overview of all SauerkrautLM-Mixtral models](#all-sauerkrautlm-mixtral-models)
351
+ 2. [Model Details](#model-details)
352
+ - [Prompt template](#prompt-template)
353
+ - [Training Dataset](#training-dataset)
354
+ - [Data Contamination Test](#data-contamination-test-results)
355
+ 3. [Evaluation](#evaluation)
356
+ 5. [Disclaimer](#disclaimer)
357
+ 6. [Contact](#contact)
358
+ 7. [Collaborations](#collaborations)
359
+ 8. [Acknowledgement](#acknowledgement)
360
+
361
+
362
+ ## All SauerkrautLM-Mixtral Models
363
+
364
+ | Model | HF | GPTQ | GGUF | AWQ |
365
+ |-------|-------|-------|-------|-------|
366
+ | SauerkrautLM-Mixtral-8x7B-Instruct | [Link](https://huggingface.co/VAGOsolutions/SauerkrautLM-Mixtral-8x7B-Instruct) | coming soon | coming soon | coming soon |
367
+ | SauerkrautLM-Mixtral-8x7B | [Link](https://huggingface.co/VAGOsolutions/SauerkrautLM-Mixtral-8x7B) | coming soon | coming soon | coming soon |
368
+
369
+ ## Model Details
370
+ **SauerkrautLM-Mixtral-8x7B-Instruct**
371
+ - **Model Type:** SauerkrautLM-Mixtral-8x7B-Instruct-v0.1 is a Mixture of Experts (MoE) Model based on [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
372
+ - **Language(s):** English, German, French, Italian, Spanish
373
+ - **License:** APACHE 2.0
374
+ - **Contact:** [Website](https://vago-solutions.de/#Kontakt) [David Golchinfar](mailto:[email protected])
375
+
376
+ ### Training Dataset:
377
+
378
+ SauerkrautLM-Mixtral-8x7B-Instruct was trained with mix of German data augmentation and translated data.
379
+ Aligned through **DPO** with our **new German SauerkrautLM-DPO dataset** based on parts of the SFT SauerkrautLM dataset
380
+ as chosen answers and [Sauerkraut-7b-HerO](https://huggingface.co/VAGOsolutions/SauerkrautLM-7b-HerO) as rejected answers. Added with additional **translated Parts of the [HuggingFaceH4/ultrafeedback_binarized](https://huggingface.co/datasets/HuggingFaceH4/ultrafeedback_binarized)** (Our dataset do not contain any TruthfulQA prompts - check Data Contamination Test Results) and **[argilla/distilabel-math-preference-dpo](https://huggingface.co/datasets/argilla/distilabel-math-preference-dpo).**
381
+ We found, that only a simple translation of training data can lead to unnatural German phrasings.
382
+ Data augmentation techniques were used to grant grammatical, syntactical correctness and a more natural German wording in our training data.
383
+
384
+ ### Data Contamination Test Results
385
+
386
+ Some models on the HuggingFace leaderboard had problems with wrong data getting mixed in.
387
+ We checked our SauerkrautLM-DPO dataset with a special test [1] on a smaller model for this problem.
388
+ The HuggingFace team used the same methods [2, 3].
389
+
390
+ Our results, with `result < 0.1, %:` being well below 0.9, indicate that our dataset is free from contamination.
391
+
392
+ *The data contamination test results of HellaSwag and Winograde will be added once [1] supports them.*
393
+
394
+ | Dataset | ARC | MMLU | TruthfulQA | GSM8K |
395
+ |------------------------------|-------|-------|-------|-------|
396
+ | **SauerkrautLM-DPO**| result < 0.1, %: 0.0 |result < 0.1, %: 0.09 | result < 0.1, %: 0.13 | result < 0.1, %: 0.16 |
397
+
398
+ [1] https://github.com/swj0419/detect-pretrain-code-contamination
399
+
400
+ [2] https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard/discussions/474#657f2245365456e362412a06
401
+
402
+ [3] https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard/discussions/265#657b6debf81f6b44b8966230
403
+
404
+ ### Prompt Template:
405
+ ```
406
+ [INST] Instruction [/INST] Model answer [INST] Follow-up instruction [/INST]
407
+ ```
408
+ ## Evaluation
409
+ ![Harness](https://vago-solutions.de/wp-content/uploads/2023/12/MOE_Instruct.png "SauerkrautLM-Mixtral-8x7B-Instruct Harness")
410
+ *evaluated with lm-evaluation-harness v0.3.0 - mmlu coming soon
411
+
412
+ *All benchmarks were performed with a sliding window of 4096. New Benchmarks with Sliding Window null coming soon
413
+
414
+ ## Disclaimer
415
+ We must inform users that despite our best efforts in data cleansing, the possibility of uncensored content slipping through cannot be entirely ruled out.
416
+ However, we cannot guarantee consistently appropriate behavior. Therefore, if you encounter any issues or come across inappropriate content, we kindly request that you inform us through the contact information provided.
417
+ Additionally, it is essential to understand that the licensing of these models does not constitute legal advice. We are not held responsible for the actions of third parties who utilize our models. These models may be employed for commercial purposes, and the Apache 2.0 remains applicable and is included with the model files.
418
+  
419
+ ## Contact
420
+ If you are interested in customized LLMs for business applications, please get in contact with us via our website or contact us at [Dr. Daryoush Vaziri](mailto:[email protected]). We are also grateful for your feedback and suggestions.
421
+  
422
+ ## Collaborations
423
+ We are also keenly seeking support and investment for our startup, VAGO solutions, where we continuously advance the development of robust language models designed to address a diverse range of purposes and requirements. If the prospect of collaboratively navigating future challenges excites you, we warmly invite you to reach out to us.
424
+
425
+ ## Acknowledgement
426
+ Many thanks to [argilla](https://huggingface.co/datasets/argilla) and [Huggingface](https://huggingface.co) for providing such valuable datasets to the Open-Source community. And of course a big thanks to MistralAI for providing the open source community with their latest technology!