aashish1904
commited on
Commit
β’
2e813da
1
Parent(s):
a15cdf6
Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -1,4 +1,6 @@
|
|
|
|
1 |
---
|
|
|
2 |
language:
|
3 |
- en
|
4 |
license: other
|
@@ -17,7 +19,7 @@ tags:
|
|
17 |
- math
|
18 |
- llama
|
19 |
- llama3
|
20 |
-
base_model:
|
21 |
datasets:
|
22 |
- allenai/ai2_arc
|
23 |
- camel-ai/physics
|
@@ -75,8 +77,7 @@ model-index:
|
|
75 |
value: 62.46
|
76 |
name: normalized accuracy
|
77 |
source:
|
78 |
-
url:
|
79 |
-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
80 |
name: Open LLM Leaderboard
|
81 |
- task:
|
82 |
type: text-generation
|
@@ -92,8 +93,7 @@ model-index:
|
|
92 |
value: 82.41
|
93 |
name: normalized accuracy
|
94 |
source:
|
95 |
-
url:
|
96 |
-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
97 |
name: Open LLM Leaderboard
|
98 |
- task:
|
99 |
type: text-generation
|
@@ -110,8 +110,7 @@ model-index:
|
|
110 |
value: 66.19
|
111 |
name: accuracy
|
112 |
source:
|
113 |
-
url:
|
114 |
-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
115 |
name: Open LLM Leaderboard
|
116 |
- task:
|
117 |
type: text-generation
|
@@ -127,8 +126,7 @@ model-index:
|
|
127 |
- type: mc2
|
128 |
value: 55.1
|
129 |
source:
|
130 |
-
url:
|
131 |
-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
132 |
name: Open LLM Leaderboard
|
133 |
- task:
|
134 |
type: text-generation
|
@@ -145,8 +143,7 @@ model-index:
|
|
145 |
value: 79.32
|
146 |
name: accuracy
|
147 |
source:
|
148 |
-
url:
|
149 |
-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
150 |
name: Open LLM Leaderboard
|
151 |
- task:
|
152 |
type: text-generation
|
@@ -163,17 +160,114 @@ model-index:
|
|
163 |
value: 66.11
|
164 |
name: accuracy
|
165 |
source:
|
166 |
-
url:
|
167 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
168 |
name: Open LLM Leaderboard
|
169 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
170 |
---
|
171 |
|
172 |
-
|
|
|
|
|
|
|
|
|
173 |
|
174 |
-
|
175 |
|
176 |
-
|
|
|
|
|
177 |
|
178 |
This model is a full fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on diverse datasets.
|
179 |
|
@@ -372,6 +466,19 @@ The results of this filtering process and its outcomes are in the data folder of
|
|
372 |
|
373 |
[Weyaxi/Einstein-v6.1-Llama3-8B/data](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B/tree/main/data)
|
374 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
375 |
|
376 |
# π― [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
377 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
|
@@ -386,6 +493,38 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
|
|
386 |
|Winogrande (5-shot) |79.32|
|
387 |
|GSM8k (5-shot) |66.11|
|
388 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
389 |
# π€ Additional information about training
|
390 |
|
391 |
This model is full fine-tuned for 2 epoch.
|
@@ -396,4 +535,20 @@ Total number of steps was 2026.
|
|
396 |
|
397 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/Ycs7ZpoqmxFt0u9rybCO1.png)
|
398 |
|
399 |
-
</details><br>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
|
2 |
---
|
3 |
+
|
4 |
language:
|
5 |
- en
|
6 |
license: other
|
|
|
19 |
- math
|
20 |
- llama
|
21 |
- llama3
|
22 |
+
base_model: meta-llama/Meta-Llama-3-8B
|
23 |
datasets:
|
24 |
- allenai/ai2_arc
|
25 |
- camel-ai/physics
|
|
|
77 |
value: 62.46
|
78 |
name: normalized accuracy
|
79 |
source:
|
80 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
|
|
81 |
name: Open LLM Leaderboard
|
82 |
- task:
|
83 |
type: text-generation
|
|
|
93 |
value: 82.41
|
94 |
name: normalized accuracy
|
95 |
source:
|
96 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
|
|
97 |
name: Open LLM Leaderboard
|
98 |
- task:
|
99 |
type: text-generation
|
|
|
110 |
value: 66.19
|
111 |
name: accuracy
|
112 |
source:
|
113 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
|
|
114 |
name: Open LLM Leaderboard
|
115 |
- task:
|
116 |
type: text-generation
|
|
|
126 |
- type: mc2
|
127 |
value: 55.1
|
128 |
source:
|
129 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
|
|
130 |
name: Open LLM Leaderboard
|
131 |
- task:
|
132 |
type: text-generation
|
|
|
143 |
value: 79.32
|
144 |
name: accuracy
|
145 |
source:
|
146 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
|
|
147 |
name: Open LLM Leaderboard
|
148 |
- task:
|
149 |
type: text-generation
|
|
|
160 |
value: 66.11
|
161 |
name: accuracy
|
162 |
source:
|
163 |
+
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
164 |
+
name: Open LLM Leaderboard
|
165 |
+
- task:
|
166 |
+
type: text-generation
|
167 |
+
name: Text Generation
|
168 |
+
dataset:
|
169 |
+
name: IFEval (0-Shot)
|
170 |
+
type: HuggingFaceH4/ifeval
|
171 |
+
args:
|
172 |
+
num_few_shot: 0
|
173 |
+
metrics:
|
174 |
+
- type: inst_level_strict_acc and prompt_level_strict_acc
|
175 |
+
value: 45.68
|
176 |
+
name: strict accuracy
|
177 |
+
source:
|
178 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
179 |
+
name: Open LLM Leaderboard
|
180 |
+
- task:
|
181 |
+
type: text-generation
|
182 |
+
name: Text Generation
|
183 |
+
dataset:
|
184 |
+
name: BBH (3-Shot)
|
185 |
+
type: BBH
|
186 |
+
args:
|
187 |
+
num_few_shot: 3
|
188 |
+
metrics:
|
189 |
+
- type: acc_norm
|
190 |
+
value: 29.38
|
191 |
+
name: normalized accuracy
|
192 |
+
source:
|
193 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
194 |
+
name: Open LLM Leaderboard
|
195 |
+
- task:
|
196 |
+
type: text-generation
|
197 |
+
name: Text Generation
|
198 |
+
dataset:
|
199 |
+
name: MATH Lvl 5 (4-Shot)
|
200 |
+
type: hendrycks/competition_math
|
201 |
+
args:
|
202 |
+
num_few_shot: 4
|
203 |
+
metrics:
|
204 |
+
- type: exact_match
|
205 |
+
value: 5.74
|
206 |
+
name: exact match
|
207 |
+
source:
|
208 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
209 |
+
name: Open LLM Leaderboard
|
210 |
+
- task:
|
211 |
+
type: text-generation
|
212 |
+
name: Text Generation
|
213 |
+
dataset:
|
214 |
+
name: GPQA (0-shot)
|
215 |
+
type: Idavidrein/gpqa
|
216 |
+
args:
|
217 |
+
num_few_shot: 0
|
218 |
+
metrics:
|
219 |
+
- type: acc_norm
|
220 |
+
value: 4.25
|
221 |
+
name: acc_norm
|
222 |
+
source:
|
223 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
224 |
name: Open LLM Leaderboard
|
225 |
+
- task:
|
226 |
+
type: text-generation
|
227 |
+
name: Text Generation
|
228 |
+
dataset:
|
229 |
+
name: MuSR (0-shot)
|
230 |
+
type: TAUR-Lab/MuSR
|
231 |
+
args:
|
232 |
+
num_few_shot: 0
|
233 |
+
metrics:
|
234 |
+
- type: acc_norm
|
235 |
+
value: 11.23
|
236 |
+
name: acc_norm
|
237 |
+
source:
|
238 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
239 |
+
name: Open LLM Leaderboard
|
240 |
+
- task:
|
241 |
+
type: text-generation
|
242 |
+
name: Text Generation
|
243 |
+
dataset:
|
244 |
+
name: MMLU-PRO (5-shot)
|
245 |
+
type: TIGER-Lab/MMLU-Pro
|
246 |
+
config: main
|
247 |
+
split: test
|
248 |
+
args:
|
249 |
+
num_few_shot: 5
|
250 |
+
metrics:
|
251 |
+
- type: acc
|
252 |
+
value: 23.68
|
253 |
+
name: accuracy
|
254 |
+
source:
|
255 |
+
url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
|
256 |
+
name: Open LLM Leaderboard
|
257 |
+
|
258 |
---
|
259 |
|
260 |
+
[![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
|
261 |
+
|
262 |
+
|
263 |
+
# QuantFactory/Einstein-v6.1-Llama3-8B-GGUF
|
264 |
+
This is quantized version of [Weyaxi/Einstein-v6.1-Llama3-8B](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) created using llama.cpp
|
265 |
|
266 |
+
# Original Model Card
|
267 |
|
268 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/5s12oq859qLfDkkTNam_C.png)
|
269 |
+
|
270 |
+
# π¬ Einstein-v6.1-Llama3-8B
|
271 |
|
272 |
This model is a full fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on diverse datasets.
|
273 |
|
|
|
466 |
|
467 |
[Weyaxi/Einstein-v6.1-Llama3-8B/data](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B/tree/main/data)
|
468 |
|
469 |
+
# π Quantizationed versions
|
470 |
+
|
471 |
+
## GGUF [@bartowski](https://huggingface.co/bartowski)
|
472 |
+
|
473 |
+
- https://huggingface.co/bartowski/Einstein-v6.1-Llama3-8B-GGUF
|
474 |
+
|
475 |
+
## ExLlamaV2 [@bartowski](https://huggingface.co/bartowski)
|
476 |
+
|
477 |
+
- https://huggingface.co/bartowski/Einstein-v6.1-Llama3-8B-exl2
|
478 |
+
|
479 |
+
## AWQ [@solidrust](https://huggingface.co/solidrust)
|
480 |
+
|
481 |
+
- https://huggingface.co/solidrust/Einstein-v6.1-Llama3-8B-AWQ
|
482 |
|
483 |
# π― [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
484 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
|
|
|
493 |
|Winogrande (5-shot) |79.32|
|
494 |
|GSM8k (5-shot) |66.11|
|
495 |
|
496 |
+
# π― [Open LLM Leaderboard v2 Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
|
497 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
|
498 |
+
|
499 |
+
| Metric |Value|
|
500 |
+
|-------------------|----:|
|
501 |
+
|Avg. |19.99|
|
502 |
+
|IFEval (0-Shot) |45.68|
|
503 |
+
|BBH (3-Shot) |29.38|
|
504 |
+
|MATH Lvl 5 (4-Shot)| 5.74|
|
505 |
+
|GPQA (0-shot) | 4.25|
|
506 |
+
|MuSR (0-shot) |11.23|
|
507 |
+
|MMLU-PRO (5-shot) |23.68|
|
508 |
+
|
509 |
+
|
510 |
+
# π Some resources, discussions and reviews aboout this model
|
511 |
+
|
512 |
+
#### π¦ Announcement tweet:
|
513 |
+
|
514 |
+
- https://twitter.com/Weyaxi/status/1783050724659675627
|
515 |
+
|
516 |
+
#### π Reddit post in r/LocalLLaMA:
|
517 |
+
|
518 |
+
- https://www.reddit.com/r/LocalLLaMA/comments/1cdlym1/introducing_einstein_v61_based_on_the_new_llama3/
|
519 |
+
|
520 |
+
#### βΆοΈ Youtube Video(s)
|
521 |
+
|
522 |
+
- [Install Einstein v6.1 Llama3-8B Locally on Windows](https://www.youtube.com/watch?v=VePvv6OM0JY)
|
523 |
+
|
524 |
+
#### π± Octopus-V4-3B
|
525 |
+
|
526 |
+
- [Octopus-V4-3B](https://huggingface.co/NexaAIDev/Octopus-v4) leverages the incredible physics capabilities of [Einstein-v6.1-Llama3-8B](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) in their model.
|
527 |
+
|
528 |
# π€ Additional information about training
|
529 |
|
530 |
This model is full fine-tuned for 2 epoch.
|
|
|
535 |
|
536 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/Ycs7ZpoqmxFt0u9rybCO1.png)
|
537 |
|
538 |
+
</details><br>
|
539 |
+
|
540 |
+
# π€ Acknowledgments
|
541 |
+
|
542 |
+
Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
|
543 |
+
|
544 |
+
Thanks to all the dataset authors mentioned in the datasets section.
|
545 |
+
|
546 |
+
Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
|
547 |
+
|
548 |
+
Thanks to all open source AI community.
|
549 |
+
|
550 |
+
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
551 |
+
|
552 |
+
If you would like to support me:
|
553 |
+
|
554 |
+
[β Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)
|