srt-42 commited on
Commit
edd9799
1 Parent(s): 64c03fe

add model card

Browse files
Files changed (1) hide show
  1. README.md +228 -1
README.md CHANGED
@@ -1,3 +1,230 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-nc-4.0
3
+ tags:
4
+ - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - dpo
8
+ - rlhf
9
+ - quantized
10
+ - 4-bit
11
+ - AWQ
12
+ - text-generation
13
+ - autotrain_compatible
14
+ - endpoints_compatible
15
+ - chatml
16
+ base_model: mlabonne/Beagle14-7B
17
+ model-index:
18
+ - name: NeuralBeagle14-7B
19
+ results:
20
+ - task:
21
+ type: text-generation
22
+ name: Text Generation
23
+ dataset:
24
+ name: AI2 Reasoning Challenge (25-Shot)
25
+ type: ai2_arc
26
+ config: ARC-Challenge
27
+ split: test
28
+ args:
29
+ num_few_shot: 25
30
+ metrics:
31
+ - type: acc_norm
32
+ value: 72.95
33
+ name: normalized accuracy
34
+ source:
35
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
36
+ name: Open LLM Leaderboard
37
+ - task:
38
+ type: text-generation
39
+ name: Text Generation
40
+ dataset:
41
+ name: HellaSwag (10-Shot)
42
+ type: hellaswag
43
+ split: validation
44
+ args:
45
+ num_few_shot: 10
46
+ metrics:
47
+ - type: acc_norm
48
+ value: 88.34
49
+ name: normalized accuracy
50
+ source:
51
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
52
+ name: Open LLM Leaderboard
53
+ - task:
54
+ type: text-generation
55
+ name: Text Generation
56
+ dataset:
57
+ name: MMLU (5-Shot)
58
+ type: cais/mmlu
59
+ config: all
60
+ split: test
61
+ args:
62
+ num_few_shot: 5
63
+ metrics:
64
+ - type: acc
65
+ value: 64.55
66
+ name: accuracy
67
+ source:
68
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
69
+ name: Open LLM Leaderboard
70
+ - task:
71
+ type: text-generation
72
+ name: Text Generation
73
+ dataset:
74
+ name: TruthfulQA (0-shot)
75
+ type: truthful_qa
76
+ config: multiple_choice
77
+ split: validation
78
+ args:
79
+ num_few_shot: 0
80
+ metrics:
81
+ - type: mc2
82
+ value: 69.93
83
+ source:
84
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
85
+ name: Open LLM Leaderboard
86
+ - task:
87
+ type: text-generation
88
+ name: Text Generation
89
+ dataset:
90
+ name: Winogrande (5-shot)
91
+ type: winogrande
92
+ config: winogrande_xl
93
+ split: validation
94
+ args:
95
+ num_few_shot: 5
96
+ metrics:
97
+ - type: acc
98
+ value: 82.4
99
+ name: accuracy
100
+ source:
101
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
102
+ name: Open LLM Leaderboard
103
+ - task:
104
+ type: text-generation
105
+ name: Text Generation
106
+ dataset:
107
+ name: GSM8k (5-shot)
108
+ type: gsm8k
109
+ config: main
110
+ split: test
111
+ args:
112
+ num_few_shot: 5
113
+ metrics:
114
+ - type: acc
115
+ value: 70.28
116
+ name: accuracy
117
+ source:
118
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=mlabonne/NeuralBeagle14-7B
119
+ name: Open LLM Leaderboard
120
+ library_name: transformers
121
+ language:
122
+ - en
123
+ model_creator: mlabonne
124
+ model_name: NeuralBeagle14-7B
125
+ model_type: mistral
126
+ pipeline_tag: text-generation
127
+ inference: false
128
+ prompt_template: '<|im_start|>system
129
+
130
+ {system_message}<|im_end|>
131
+
132
+ <|im_start|>user
133
+
134
+ {prompt}<|im_end|>
135
+
136
+ <|im_start|>assistant
137
+
138
+ '
139
+ quantized_by: Suparious
140
  ---
141
+ # mlabonne/NeuralBeagle14-7B AWQ
142
+
143
+ - Model creator: [mlabonne](https://huggingface.co/mlabonne)
144
+ - Original model: [NeuralBeagle14-7B](https://huggingface.co/mlabonne/NeuralBeagle14-7B)
145
+
146
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6437292ecd93f4c9a34b0d47/Ip9wEG2Ne4vihNStHSDvX.png)
147
+
148
+ ## Model Summary
149
+
150
+ NeuralBeagle14-7B is a DPO fine-tune of [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) using the [argilla/distilabel-intel-orca-dpo-pairs](https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs) preference dataset and my DPO notebook from [this article](https://towardsdatascience.com/fine-tune-a-mistral-7b-model-with-direct-preference-optimization-708042745aac).
151
+
152
+ It is based on a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
153
+ * [fblgit/UNA-TheBeagle-7b-v1](https://huggingface.co/fblgit/UNA-TheBeagle-7b-v1), based on jondurbin's [repo](https://github.com/jondurbin/bagel) and [jondurbin/bagel-v0.3](https://huggingface.co/datasets/jondurbin/bagel-v0.3])
154
+ * [argilla/distilabeled-Marcoro14-7B-slerp](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp), based on [mlabonne/Marcoro14-7B-slerp](https://huggingface.co/mlabonne/Marcoro14-7B-slerp)
155
+
156
+ Thanks [Argilla](https://huggingface.co/argilla) for providing the dataset and the training recipe [here](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp). 💪
157
+
158
+ You can try it out in this [Space](https://huggingface.co/spaces/mlabonne/NeuralBeagle14-7B-GGUF-Chat) (GGUF Q4_K_M).
159
+
160
+ ## How to use
161
+
162
+ ### Install the necessary packages
163
+
164
+ ```bash
165
+ pip install --upgrade autoawq autoawq-kernels
166
+ ```
167
+
168
+ ### Example Python code
169
+
170
+ ```python
171
+ from awq import AutoAWQForCausalLM
172
+ from transformers import AutoTokenizer, TextStreamer
173
+
174
+ model_path = "solidrust/NeuralBeagle14-7B-AWQ"
175
+ system_message = "You are Beagle, incarnated as a powerful AI."
176
+
177
+ # Load model
178
+ model = AutoAWQForCausalLM.from_quantized(model_path,
179
+ fuse_layers=True)
180
+ tokenizer = AutoTokenizer.from_pretrained(model_path,
181
+ trust_remote_code=True)
182
+ streamer = TextStreamer(tokenizer,
183
+ skip_prompt=True,
184
+ skip_special_tokens=True)
185
+
186
+ # Convert prompt to tokens
187
+ prompt_template = """\
188
+ <|im_start|>system
189
+ {system_message}<|im_end|>
190
+ <|im_start|>user
191
+ {prompt}<|im_end|>
192
+ <|im_start|>assistant"""
193
+
194
+ prompt = "You're standing on the surface of the Earth. "\
195
+ "You walk one mile south, one mile west and one mile north. "\
196
+ "You end up exactly where you started. Where are you?"
197
+
198
+ tokens = tokenizer(prompt_template.format(system_message=system_message,prompt=prompt),
199
+ return_tensors='pt').input_ids.cuda()
200
+
201
+ # Generate output
202
+ generation_output = model.generate(tokens,
203
+ streamer=streamer,
204
+ max_new_tokens=512)
205
+
206
+ ```
207
+
208
+ ### About AWQ
209
+
210
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
211
+
212
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
213
+
214
+ It is supported by:
215
+
216
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
217
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
218
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
219
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
220
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
221
+
222
+ ## Prompt template: ChatML
223
+
224
+ ```plaintext
225
+ <|im_start|>system
226
+ {system_message}<|im_end|>
227
+ <|im_start|>user
228
+ {prompt}<|im_end|>
229
+ <|im_start|>assistant
230
+ ```