|
--- |
|
license: apache-2.0 |
|
library_name: transformers |
|
tags: |
|
- merge |
|
pipeline_tag: text-generation |
|
model-index: |
|
- name: TheTop-5x7B-Instruct-T-v0.1 |
|
results: |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: AI2 Reasoning Challenge (25-Shot) |
|
type: ai2_arc |
|
config: ARC-Challenge |
|
split: test |
|
args: |
|
num_few_shot: 25 |
|
metrics: |
|
- type: acc_norm |
|
value: 73.63 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: HellaSwag (10-Shot) |
|
type: hellaswag |
|
split: validation |
|
args: |
|
num_few_shot: 10 |
|
metrics: |
|
- type: acc_norm |
|
value: 88.85 |
|
name: normalized accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: MMLU (5-Shot) |
|
type: cais/mmlu |
|
config: all |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 64.22 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: TruthfulQA (0-shot) |
|
type: truthful_qa |
|
config: multiple_choice |
|
split: validation |
|
args: |
|
num_few_shot: 0 |
|
metrics: |
|
- type: mc2 |
|
value: 70.78 |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: Winogrande (5-shot) |
|
type: winogrande |
|
config: winogrande_xl |
|
split: validation |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 85.79 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
- task: |
|
type: text-generation |
|
name: Text Generation |
|
dataset: |
|
name: GSM8k (5-shot) |
|
type: gsm8k |
|
config: main |
|
split: test |
|
args: |
|
num_few_shot: 5 |
|
metrics: |
|
- type: acc |
|
value: 66.49 |
|
name: accuracy |
|
source: |
|
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MaziyarPanahi/TheTop-5x7B-Instruct-T-v0.1 |
|
name: Open LLM Leaderboard |
|
--- |
|
|
|
Merge of top 7B models with TIES method |
|
|
|
> mergekit is a toolkit for merging pre-trained language models. mergekit uses an out-of-core approach to perform unreasonably elaborate merges in resource-constrained situations. Merges can be run entirely on CPU or accelerated with as little as 8 GB of VRAM. Many merging algorithms are supported, with more coming as they catch my attention. |
|
|
|
## Eval |
|
|
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/5fd5e18a90b6dc4633f6d292/YdjDNbmytFAPv-VGiuhx7.png) |
|
|
|
```python |
|
{ |
|
"all": { |
|
"acc": 0.6487801278765712, |
|
"acc_stderr": 0.03219011246717665, |
|
"acc_norm": 0.6479445077777353, |
|
"acc_norm_stderr": 0.032868022907407396, |
|
"mc1": 0.5862913096695227, |
|
"mc1_stderr": 0.0172408618120998, |
|
"mc2": 0.7078078883926877, |
|
"mc2_stderr": 0.015097515102384168 |
|
}, |
|
"harness|arc:challenge|25": { |
|
"acc": 0.7167235494880546, |
|
"acc_stderr": 0.013167478735134575, |
|
"acc_norm": 0.7363481228668942, |
|
"acc_norm_stderr": 0.012875929151297044 |
|
}, |
|
"harness|hellaswag|10": { |
|
"acc": 0.7321250746863175, |
|
"acc_stderr": 0.004419469983939178, |
|
"acc_norm": 0.8884684325831508, |
|
"acc_norm_stderr": 0.0031414591751392717 |
|
}, |
|
"harness|hendrycksTest-abstract_algebra|5": { |
|
"acc": 0.31, |
|
"acc_stderr": 0.04648231987117316, |
|
"acc_norm": 0.31, |
|
"acc_norm_stderr": 0.04648231987117316 |
|
}, |
|
"harness|hendrycksTest-anatomy|5": { |
|
"acc": 0.6518518518518519, |
|
"acc_stderr": 0.041153246103369526, |
|
"acc_norm": 0.6518518518518519, |
|
"acc_norm_stderr": 0.041153246103369526 |
|
}, |
|
"harness|hendrycksTest-astronomy|5": { |
|
"acc": 0.7039473684210527, |
|
"acc_stderr": 0.03715062154998904, |
|
"acc_norm": 0.7039473684210527, |
|
"acc_norm_stderr": 0.03715062154998904 |
|
}, |
|
"harness|hendrycksTest-business_ethics|5": { |
|
"acc": 0.61, |
|
"acc_stderr": 0.04902071300001975, |
|
"acc_norm": 0.61, |
|
"acc_norm_stderr": 0.04902071300001975 |
|
}, |
|
"harness|hendrycksTest-clinical_knowledge|5": { |
|
"acc": 0.7132075471698113, |
|
"acc_stderr": 0.02783491252754407, |
|
"acc_norm": 0.7132075471698113, |
|
"acc_norm_stderr": 0.02783491252754407 |
|
}, |
|
"harness|hendrycksTest-college_biology|5": { |
|
"acc": 0.75, |
|
"acc_stderr": 0.03621034121889507, |
|
"acc_norm": 0.75, |
|
"acc_norm_stderr": 0.03621034121889507 |
|
}, |
|
"harness|hendrycksTest-college_chemistry|5": { |
|
"acc": 0.46, |
|
"acc_stderr": 0.05009082659620333, |
|
"acc_norm": 0.46, |
|
"acc_norm_stderr": 0.05009082659620333 |
|
}, |
|
"harness|hendrycksTest-college_computer_science|5": { |
|
"acc": 0.55, |
|
"acc_stderr": 0.05, |
|
"acc_norm": 0.55, |
|
"acc_norm_stderr": 0.05 |
|
}, |
|
"harness|hendrycksTest-college_mathematics|5": { |
|
"acc": 0.3, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.3, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|hendrycksTest-college_medicine|5": { |
|
"acc": 0.6589595375722543, |
|
"acc_stderr": 0.036146654241808254, |
|
"acc_norm": 0.6589595375722543, |
|
"acc_norm_stderr": 0.036146654241808254 |
|
}, |
|
"harness|hendrycksTest-college_physics|5": { |
|
"acc": 0.43137254901960786, |
|
"acc_stderr": 0.04928099597287534, |
|
"acc_norm": 0.43137254901960786, |
|
"acc_norm_stderr": 0.04928099597287534 |
|
}, |
|
"harness|hendrycksTest-computer_security|5": { |
|
"acc": 0.77, |
|
"acc_stderr": 0.04229525846816506, |
|
"acc_norm": 0.77, |
|
"acc_norm_stderr": 0.04229525846816506 |
|
}, |
|
"harness|hendrycksTest-conceptual_physics|5": { |
|
"acc": 0.548936170212766, |
|
"acc_stderr": 0.032529096196131965, |
|
"acc_norm": 0.548936170212766, |
|
"acc_norm_stderr": 0.032529096196131965 |
|
}, |
|
"harness|hendrycksTest-econometrics|5": { |
|
"acc": 0.49122807017543857, |
|
"acc_stderr": 0.04702880432049615, |
|
"acc_norm": 0.49122807017543857, |
|
"acc_norm_stderr": 0.04702880432049615 |
|
}, |
|
"harness|hendrycksTest-electrical_engineering|5": { |
|
"acc": 0.5517241379310345, |
|
"acc_stderr": 0.04144311810878152, |
|
"acc_norm": 0.5517241379310345, |
|
"acc_norm_stderr": 0.04144311810878152 |
|
}, |
|
"harness|hendrycksTest-elementary_mathematics|5": { |
|
"acc": 0.4126984126984127, |
|
"acc_stderr": 0.025355741263055277, |
|
"acc_norm": 0.4126984126984127, |
|
"acc_norm_stderr": 0.025355741263055277 |
|
}, |
|
"harness|hendrycksTest-formal_logic|5": { |
|
"acc": 0.49206349206349204, |
|
"acc_stderr": 0.044715725362943486, |
|
"acc_norm": 0.49206349206349204, |
|
"acc_norm_stderr": 0.044715725362943486 |
|
}, |
|
"harness|hendrycksTest-global_facts|5": { |
|
"acc": 0.35, |
|
"acc_stderr": 0.047937248544110196, |
|
"acc_norm": 0.35, |
|
"acc_norm_stderr": 0.047937248544110196 |
|
}, |
|
"harness|hendrycksTest-high_school_biology|5": { |
|
"acc": 0.7967741935483871, |
|
"acc_stderr": 0.02289168798455496, |
|
"acc_norm": 0.7967741935483871, |
|
"acc_norm_stderr": 0.02289168798455496 |
|
}, |
|
"harness|hendrycksTest-high_school_chemistry|5": { |
|
"acc": 0.5024630541871922, |
|
"acc_stderr": 0.035179450386910616, |
|
"acc_norm": 0.5024630541871922, |
|
"acc_norm_stderr": 0.035179450386910616 |
|
}, |
|
"harness|hendrycksTest-high_school_computer_science|5": { |
|
"acc": 0.7, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.7, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|hendrycksTest-high_school_european_history|5": { |
|
"acc": 0.7575757575757576, |
|
"acc_stderr": 0.03346409881055953, |
|
"acc_norm": 0.7575757575757576, |
|
"acc_norm_stderr": 0.03346409881055953 |
|
}, |
|
"harness|hendrycksTest-high_school_geography|5": { |
|
"acc": 0.803030303030303, |
|
"acc_stderr": 0.028335609732463362, |
|
"acc_norm": 0.803030303030303, |
|
"acc_norm_stderr": 0.028335609732463362 |
|
}, |
|
"harness|hendrycksTest-high_school_government_and_politics|5": { |
|
"acc": 0.9067357512953368, |
|
"acc_stderr": 0.020986854593289733, |
|
"acc_norm": 0.9067357512953368, |
|
"acc_norm_stderr": 0.020986854593289733 |
|
}, |
|
"harness|hendrycksTest-high_school_macroeconomics|5": { |
|
"acc": 0.6487179487179487, |
|
"acc_stderr": 0.024203665177902803, |
|
"acc_norm": 0.6487179487179487, |
|
"acc_norm_stderr": 0.024203665177902803 |
|
}, |
|
"harness|hendrycksTest-high_school_mathematics|5": { |
|
"acc": 0.3333333333333333, |
|
"acc_stderr": 0.02874204090394848, |
|
"acc_norm": 0.3333333333333333, |
|
"acc_norm_stderr": 0.02874204090394848 |
|
}, |
|
"harness|hendrycksTest-high_school_microeconomics|5": { |
|
"acc": 0.6554621848739496, |
|
"acc_stderr": 0.03086868260412162, |
|
"acc_norm": 0.6554621848739496, |
|
"acc_norm_stderr": 0.03086868260412162 |
|
}, |
|
"harness|hendrycksTest-high_school_physics|5": { |
|
"acc": 0.32450331125827814, |
|
"acc_stderr": 0.038227469376587525, |
|
"acc_norm": 0.32450331125827814, |
|
"acc_norm_stderr": 0.038227469376587525 |
|
}, |
|
"harness|hendrycksTest-high_school_psychology|5": { |
|
"acc": 0.8403669724770643, |
|
"acc_stderr": 0.015703498348461763, |
|
"acc_norm": 0.8403669724770643, |
|
"acc_norm_stderr": 0.015703498348461763 |
|
}, |
|
"harness|hendrycksTest-high_school_statistics|5": { |
|
"acc": 0.5046296296296297, |
|
"acc_stderr": 0.03409825519163572, |
|
"acc_norm": 0.5046296296296297, |
|
"acc_norm_stderr": 0.03409825519163572 |
|
}, |
|
"harness|hendrycksTest-high_school_us_history|5": { |
|
"acc": 0.8235294117647058, |
|
"acc_stderr": 0.026756401538078962, |
|
"acc_norm": 0.8235294117647058, |
|
"acc_norm_stderr": 0.026756401538078962 |
|
}, |
|
"harness|hendrycksTest-high_school_world_history|5": { |
|
"acc": 0.7721518987341772, |
|
"acc_stderr": 0.02730348459906944, |
|
"acc_norm": 0.7721518987341772, |
|
"acc_norm_stderr": 0.02730348459906944 |
|
}, |
|
"harness|hendrycksTest-human_aging|5": { |
|
"acc": 0.6816143497757847, |
|
"acc_stderr": 0.03126580522513713, |
|
"acc_norm": 0.6816143497757847, |
|
"acc_norm_stderr": 0.03126580522513713 |
|
}, |
|
"harness|hendrycksTest-human_sexuality|5": { |
|
"acc": 0.7862595419847328, |
|
"acc_stderr": 0.0359546161177469, |
|
"acc_norm": 0.7862595419847328, |
|
"acc_norm_stderr": 0.0359546161177469 |
|
}, |
|
"harness|hendrycksTest-international_law|5": { |
|
"acc": 0.7851239669421488, |
|
"acc_stderr": 0.037494924487096966, |
|
"acc_norm": 0.7851239669421488, |
|
"acc_norm_stderr": 0.037494924487096966 |
|
}, |
|
"harness|hendrycksTest-jurisprudence|5": { |
|
"acc": 0.7777777777777778, |
|
"acc_stderr": 0.0401910747255735, |
|
"acc_norm": 0.7777777777777778, |
|
"acc_norm_stderr": 0.0401910747255735 |
|
}, |
|
"harness|hendrycksTest-logical_fallacies|5": { |
|
"acc": 0.7423312883435583, |
|
"acc_stderr": 0.03436150827846917, |
|
"acc_norm": 0.7423312883435583, |
|
"acc_norm_stderr": 0.03436150827846917 |
|
}, |
|
"harness|hendrycksTest-machine_learning|5": { |
|
"acc": 0.42857142857142855, |
|
"acc_stderr": 0.04697113923010212, |
|
"acc_norm": 0.42857142857142855, |
|
"acc_norm_stderr": 0.04697113923010212 |
|
}, |
|
"harness|hendrycksTest-management|5": { |
|
"acc": 0.7475728155339806, |
|
"acc_stderr": 0.04301250399690878, |
|
"acc_norm": 0.7475728155339806, |
|
"acc_norm_stderr": 0.04301250399690878 |
|
}, |
|
"harness|hendrycksTest-marketing|5": { |
|
"acc": 0.8846153846153846, |
|
"acc_stderr": 0.02093019318517933, |
|
"acc_norm": 0.8846153846153846, |
|
"acc_norm_stderr": 0.02093019318517933 |
|
}, |
|
"harness|hendrycksTest-medical_genetics|5": { |
|
"acc": 0.7, |
|
"acc_stderr": 0.046056618647183814, |
|
"acc_norm": 0.7, |
|
"acc_norm_stderr": 0.046056618647183814 |
|
}, |
|
"harness|hendrycksTest-miscellaneous|5": { |
|
"acc": 0.80970625798212, |
|
"acc_stderr": 0.014036945850381396, |
|
"acc_norm": 0.80970625798212, |
|
"acc_norm_stderr": 0.014036945850381396 |
|
}, |
|
"harness|hendrycksTest-moral_disputes|5": { |
|
"acc": 0.7369942196531792, |
|
"acc_stderr": 0.023703099525258172, |
|
"acc_norm": 0.7369942196531792, |
|
"acc_norm_stderr": 0.023703099525258172 |
|
}, |
|
"harness|hendrycksTest-moral_scenarios|5": { |
|
"acc": 0.47150837988826816, |
|
"acc_stderr": 0.016695329746015796, |
|
"acc_norm": 0.47150837988826816, |
|
"acc_norm_stderr": 0.016695329746015796 |
|
}, |
|
"harness|hendrycksTest-nutrition|5": { |
|
"acc": 0.7189542483660131, |
|
"acc_stderr": 0.025738854797818733, |
|
"acc_norm": 0.7189542483660131, |
|
"acc_norm_stderr": 0.025738854797818733 |
|
}, |
|
"harness|hendrycksTest-philosophy|5": { |
|
"acc": 0.7170418006430869, |
|
"acc_stderr": 0.025583062489984813, |
|
"acc_norm": 0.7170418006430869, |
|
"acc_norm_stderr": 0.025583062489984813 |
|
}, |
|
"harness|hendrycksTest-prehistory|5": { |
|
"acc": 0.7407407407407407, |
|
"acc_stderr": 0.024383665531035457, |
|
"acc_norm": 0.7407407407407407, |
|
"acc_norm_stderr": 0.024383665531035457 |
|
}, |
|
"harness|hendrycksTest-professional_accounting|5": { |
|
"acc": 0.475177304964539, |
|
"acc_stderr": 0.029790719243829727, |
|
"acc_norm": 0.475177304964539, |
|
"acc_norm_stderr": 0.029790719243829727 |
|
}, |
|
"harness|hendrycksTest-professional_law|5": { |
|
"acc": 0.470013037809648, |
|
"acc_stderr": 0.01274724896707906, |
|
"acc_norm": 0.470013037809648, |
|
"acc_norm_stderr": 0.01274724896707906 |
|
}, |
|
"harness|hendrycksTest-professional_medicine|5": { |
|
"acc": 0.6691176470588235, |
|
"acc_stderr": 0.028582709753898445, |
|
"acc_norm": 0.6691176470588235, |
|
"acc_norm_stderr": 0.028582709753898445 |
|
}, |
|
"harness|hendrycksTest-professional_psychology|5": { |
|
"acc": 0.6584967320261438, |
|
"acc_stderr": 0.019184639328092487, |
|
"acc_norm": 0.6584967320261438, |
|
"acc_norm_stderr": 0.019184639328092487 |
|
}, |
|
"harness|hendrycksTest-public_relations|5": { |
|
"acc": 0.6818181818181818, |
|
"acc_stderr": 0.044612721759105085, |
|
"acc_norm": 0.6818181818181818, |
|
"acc_norm_stderr": 0.044612721759105085 |
|
}, |
|
"harness|hendrycksTest-security_studies|5": { |
|
"acc": 0.7306122448979592, |
|
"acc_stderr": 0.02840125202902294, |
|
"acc_norm": 0.7306122448979592, |
|
"acc_norm_stderr": 0.02840125202902294 |
|
}, |
|
"harness|hendrycksTest-sociology|5": { |
|
"acc": 0.835820895522388, |
|
"acc_stderr": 0.026193923544454125, |
|
"acc_norm": 0.835820895522388, |
|
"acc_norm_stderr": 0.026193923544454125 |
|
}, |
|
"harness|hendrycksTest-us_foreign_policy|5": { |
|
"acc": 0.85, |
|
"acc_stderr": 0.03588702812826371, |
|
"acc_norm": 0.85, |
|
"acc_norm_stderr": 0.03588702812826371 |
|
}, |
|
"harness|hendrycksTest-virology|5": { |
|
"acc": 0.5542168674698795, |
|
"acc_stderr": 0.03869543323472101, |
|
"acc_norm": 0.5542168674698795, |
|
"acc_norm_stderr": 0.03869543323472101 |
|
}, |
|
"harness|hendrycksTest-world_religions|5": { |
|
"acc": 0.8245614035087719, |
|
"acc_stderr": 0.029170885500727665, |
|
"acc_norm": 0.8245614035087719, |
|
"acc_norm_stderr": 0.029170885500727665 |
|
}, |
|
"harness|truthfulqa:mc|0": { |
|
"mc1": 0.5862913096695227, |
|
"mc1_stderr": 0.0172408618120998, |
|
"mc2": 0.7078078883926877, |
|
"mc2_stderr": 0.015097515102384168 |
|
}, |
|
"harness|winogrande|5": { |
|
"acc": 0.8579321231254933, |
|
"acc_stderr": 0.009812000391679367 |
|
}, |
|
"harness|gsm8k|5": { |
|
"acc": 0.6648976497346475, |
|
"acc_stderr": 0.013001948176422954 |
|
} |
|
} |
|
``` |
|
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) |
|
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_MaziyarPanahi__TheTop-5x7B-Instruct-T-v0.1) |
|
|
|
| Metric |Value| |
|
|---------------------------------|----:| |
|
|Avg. |74.96| |
|
|AI2 Reasoning Challenge (25-Shot)|73.63| |
|
|HellaSwag (10-Shot) |88.85| |
|
|MMLU (5-Shot) |64.22| |
|
|TruthfulQA (0-shot) |70.78| |
|
|Winogrande (5-shot) |85.79| |
|
|GSM8k (5-shot) |66.49| |
|
|
|
|