Update README.md
Browse files
README.md
CHANGED
@@ -130,18 +130,18 @@ This instruction model was built via parameter-efficient QLoRA finetuning of [ll
|
|
130 |
* Model license: Llama 2 Community License Agreement
|
131 |
* Basic usage: [notebook](assets/basic_inference_llama_2_dolphin.ipynb)
|
132 |
|
133 |
-
## Open LLM Leaderboard Evaluation Results
|
134 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
135 |
|
136 |
-
|
|
137 |
-
|
138 |
-
|
|
139 |
-
|
|
140 |
-
|
|
141 |
-
|
|
142 |
-
|
|
143 |
-
|
|
144 |
-
|
|
145 |
|
146 |
## Example prompts and responses
|
147 |
|
@@ -359,17 +359,3 @@ The license on this model does not constitute legal advice. We are not responsib
|
|
359 |
|
360 |
- PEFT 0.5.0.dev0
|
361 |
|
362 |
-
|
363 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
364 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
365 |
-
|
366 |
-
| Metric |Value|
|
367 |
-
|---------------------------------|----:|
|
368 |
-
|Avg. |65.72|
|
369 |
-
|AI2 Reasoning Challenge (25-Shot)|69.62|
|
370 |
-
|HellaSwag (10-Shot) |86.82|
|
371 |
-
|MMLU (5-Shot) |69.18|
|
372 |
-
|TruthfulQA (0-shot) |57.43|
|
373 |
-
|Winogrande (5-shot) |83.90|
|
374 |
-
|GSM8k (5-shot) |27.37|
|
375 |
-
|
|
|
130 |
* Model license: Llama 2 Community License Agreement
|
131 |
* Basic usage: [notebook](assets/basic_inference_llama_2_dolphin.ipynb)
|
132 |
|
133 |
+
## [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
134 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_dfurman__llama-2-70b-dolphin-peft)
|
135 |
|
136 |
+
| Metric |Value|
|
137 |
+
|---------------------------------|----:|
|
138 |
+
|Avg. |65.72|
|
139 |
+
|AI2 Reasoning Challenge (25-Shot)|69.62|
|
140 |
+
|HellaSwag (10-Shot) |86.82|
|
141 |
+
|MMLU (5-Shot) |69.18|
|
142 |
+
|TruthfulQA (0-shot) |57.43|
|
143 |
+
|Winogrande (5-shot) |83.90|
|
144 |
+
|GSM8k (5-shot) |27.37|
|
145 |
|
146 |
## Example prompts and responses
|
147 |
|
|
|
359 |
|
360 |
- PEFT 0.5.0.dev0
|
361 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|