model
stringlengths
4
89
revision
stringclasses
1 value
model_sha
stringlengths
0
40
results
dict
commit
stringlengths
40
40
date
timestamp[ns]
score
float64
21.8
83
h2oai/h2ogpt-research-oasst1-llama-65b
main
a6d8676aaa2ca2c25ea99180b538f0369dc70185
{ "arc:challenge": 64.8, "hellaswag": 85.9, "hendrycksTest": 63.6, "truthfulqa:mc": 48.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
65.8
h2oai/h2ogpt-oig-oasst1-256-6_9b
main
f1c9bac89b74d3487cb092788ce828fb9520c1a7
{ "arc:challenge": 39.9, "hellaswag": 65.4, "hendrycksTest": 26.4, "truthfulqa:mc": 35 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
41.7
h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt
main
754e0c90ed5d9241fdfd5a188572b3ea2152eaa7
{ "arc:challenge": 34, "hellaswag": 50.5, "hendrycksTest": 24.7, "truthfulqa:mc": 41.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
37.8
h2oai/h2ogpt-research-oig-oasst1-512-30b
main
4215e83b9038c9e61d979cf5223b29f860bace9b
{ "arc:challenge": 59, "hellaswag": 82.6, "hendrycksTest": 50.7, "truthfulqa:mc": 48.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
60.2
h2oai/h2ogpt-gm-oasst1-en-1024-20b
main
1a5b8d25587eab67d837621a6c9423e7ef6df289
{ "arc:challenge": 48, "hellaswag": 72.8, "hendrycksTest": 26, "truthfulqa:mc": 39.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
46.7
h2oai/h2ogpt-oasst1-512-12b
main
c6bb0fe363e0105839d34ca757793b61c9606f95
{ "arc:challenge": 42.3, "hellaswag": 70.2, "hendrycksTest": 26, "truthfulqa:mc": 36.4 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
43.7
h2oai/h2ogpt-gm-oasst1-en-1024-12b
main
e547fffafb382fd39ef5de35ba3b5afc1b43e74d
{ "arc:challenge": 43.1, "hellaswag": 69.7, "hendrycksTest": 25.9, "truthfulqa:mc": 38 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
44.2
minlik/chinese-alpaca-33b-merged
main
fc2535104c0b48afc42575f9fe10bbcbb7612ec3
{ "arc:challenge": 59.3, "hellaswag": 78.4, "hendrycksTest": 57.7, "truthfulqa:mc": 52.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
62
xzuyn/Alpacino-SuperCOT-13B
main
3a82b04684fe99d59556421c3f96a187049a3cec
{ "arc:challenge": 58.4, "hellaswag": 81.7, "hendrycksTest": 47.9, "truthfulqa:mc": 45.4 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.4
xzuyn/MedicWizard-7B
main
0b3ef975fb5e8ac1eae775160ab54c98221889df
{ "arc:challenge": 53.5, "hellaswag": 78.4, "hendrycksTest": 44.6, "truthfulqa:mc": 41.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
54.4
teknium/OpenHermes-2.5-Mistral-7B
main
2a54cad766bc90828354db5c4199795aecfd0df1
{ "arc:challenge": 64.9, "hellaswag": 84.3, "hendrycksTest": 63.8, "truthfulqa:mc": 52.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
66.3
teknium/OpenHermes-2-Mistral-7B
main
843a9bb94fac7d7bfc1b7c9f201efba295b6f5d6
{ "arc:challenge": 63.1, "hellaswag": 83.8, "hendrycksTest": 63.5, "truthfulqa:mc": 50.2 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
65.2
teknium/OpenHermes-13B
main
f09d0fe655ad57cce9179b7b40ea6f81e07db18c
{ "arc:challenge": 59.8, "hellaswag": 82.2, "hendrycksTest": 56.3, "truthfulqa:mc": 46 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61.1
teknium/CollectiveCognition-v1.1-Mistral-7B
main
5f57f70ec99450c70da2540e94dd7fd67be4b23c
{ "arc:challenge": 62.1, "hellaswag": 84.2, "hendrycksTest": 62.3, "truthfulqa:mc": 57.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
66.6
nathan0/mpt_delta_tuned_model_v2
main
6adb4cb4ba37f4ce9e9c3051d343addf1098182c
{ "arc:challenge": 50.7, "hellaswag": 76.4, "hendrycksTest": 28.7, "truthfulqa:mc": 35.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
47.8
nathan0/mpt_delta_tuned_model_v3
main
6adb4cb4ba37f4ce9e9c3051d343addf1098182c
{ "arc:challenge": 50.7, "hellaswag": 76.4, "hendrycksTest": 28.7, "truthfulqa:mc": 35.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
47.8
chargoddard/duplicitous-mammal-13b
main
a05d0562b8da2ac2e76aa65984e8063249bc85c8
{ "arc:challenge": 61.7, "hellaswag": 83.8, "hendrycksTest": 57.5, "truthfulqa:mc": 52.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
63.8
chargoddard/llama2-22b-blocktriangular
main
40a51343ae776b5cb39f2b4343ae8f9b676ffd58
{ "arc:challenge": 58.5, "hellaswag": 82.6, "hendrycksTest": 54.6, "truthfulqa:mc": 39.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.8
chargoddard/MelangeB-70b
main
08239fb1e30b1e42b14370f23e942bc51e76027c
{ "arc:challenge": 71.7, "hellaswag": 87.5, "hendrycksTest": 70, "truthfulqa:mc": 59.4 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
72.1
chargoddard/ypotryll-22b-epoch2-qlora
main
26fdd8fa420d72ed835c7d17086f0441db0985d4
{ "arc:challenge": 59.2, "hellaswag": 80.7, "hendrycksTest": 54.5, "truthfulqa:mc": 40.4 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.7
chargoddard/llama-2-26b-trenchcoat-stack
main
075d67c3223f4b379ab7f997c3787cd0630d80f7
{ "arc:challenge": 55, "hellaswag": 79.9, "hendrycksTest": 53.7, "truthfulqa:mc": 40.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
57.3
chargoddard/llama2-22b
main
2bece0787009b4b584f49d0e0d1b49ecf4a52da9
{ "arc:challenge": 58.5, "hellaswag": 82.6, "hendrycksTest": 54.7, "truthfulqa:mc": 39.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.9
chargoddard/platypus-2-22b-relora
main
15bca3e9b25cc2f280fec21686ef3bc445217503
{ "arc:challenge": 57.7, "hellaswag": 82.4, "hendrycksTest": 55.3, "truthfulqa:mc": 43.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59.8
chargoddard/platypus2-22b-relora
main
15bca3e9b25cc2f280fec21686ef3bc445217503
{ "arc:challenge": 57.5, "hellaswag": 82.4, "hendrycksTest": 54.9, "truthfulqa:mc": 43.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59.6
chargoddard/storytime-13b
main
233568319a636b6a7b02a4def2c51d08a3e0fbfc
{ "arc:challenge": 62, "hellaswag": 84, "hendrycksTest": 57.5, "truthfulqa:mc": 52.5 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
64
chargoddard/Chronorctypus-Limarobormes-13b
main
75c1bf5f4b40cf61873ff6487ccd3efc4f684330
{ "arc:challenge": 59.9, "hellaswag": 82.8, "hendrycksTest": 58.4, "truthfulqa:mc": 51.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
63.2
chargoddard/MelangeC-70b
main
e54a2b924dec135f3fa2373933ab8485178cde1b
{ "arc:challenge": 71.7, "hellaswag": 87.6, "hendrycksTest": 70.4, "truthfulqa:mc": 58.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
72
Open-Orca/Mistral-7B-SlimOrca
main
a9744d8cf9ce4230678a891bcf8bba7cbc0aaece
{ "arc:challenge": 62.5, "hellaswag": 83.9, "hendrycksTest": 62.8, "truthfulqa:mc": 54.2 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
65.8
Open-Orca/LlongOrca-13B-16k
main
8ea1fb205553cadbc90069d80a7e58281b6281c3
{ "arc:challenge": 62.5, "hellaswag": 82.8, "hendrycksTest": 55.5, "truthfulqa:mc": 50.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
62.7
Open-Orca/OpenOrca-Preview1-13B
main
4c558283a98348383460939afda9cb5c54544c8f
{ "arc:challenge": 54.9, "hellaswag": 78.2, "hendrycksTest": 50.1, "truthfulqa:mc": 49.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.1
Open-Orca/OpenOrca-Platypus2-13B
main
e7a40134f7eb687c6ab66d445dc7251257f8d391
{ "arc:challenge": 62.8, "hellaswag": 83.2, "hendrycksTest": 59.4, "truthfulqa:mc": 53.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
64.6
Open-Orca/LlongOrca-7B-16k
main
1370c7c595e6c8394e6332bc535ae25e21def85b
{ "arc:challenge": 57.5, "hellaswag": 79.4, "hendrycksTest": 49.4, "truthfulqa:mc": 49.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59
grantprice/Cerebras-GPT-590M-finetuned-DND
main
a0a2fbe342cdc86433913ba5f96978e4703ff672
{ "arc:challenge": 24.7, "hellaswag": 27.8, "hendrycksTest": 23.1, "truthfulqa:mc": 48.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
31
lgaalves/tinyllama-1.1b-chat-v0.3_platypus
main
0bb6ebe1d41d394bae0ed9107ec8d776d9d76a68
{ "arc:challenge": 30.3, "hellaswag": 55.1, "hendrycksTest": 26.1, "truthfulqa:mc": 39.2 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
37.7
lgaalves/gpt2_guanaco-dolly-platypus
main
6bf0a8146cf255c829ec2ad83926c8b80945b431
{ "arc:challenge": 23.5, "hellaswag": 31, "hendrycksTest": 26.4, "truthfulqa:mc": 40 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
30.2
lgaalves/mistral-7b_open_platypus
main
b9a60b9ad0fe06bd314ffe99d543f1df6ecd10da
{ "arc:challenge": 55.8, "hellaswag": 82.1, "hendrycksTest": 59.8, "truthfulqa:mc": 48.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61.6
lgaalves/gpt2_camel_physics-platypus
main
66165ff32ed8de6c39f3524a810f5e97ba6d3347
{ "arc:challenge": 23, "hellaswag": 31.3, "hendrycksTest": 26.9, "truthfulqa:mc": 39.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
30.2
lgaalves/gpt2-xl_lima
main
f7db5b1db521abd7578b95138e737637e0037ca5
{ "arc:challenge": 31.1, "hellaswag": 51.3, "hendrycksTest": 25.4, "truthfulqa:mc": 38.7 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
36.6
lgaalves/llama-2-7b-hf_open-platypus
main
c7e776f3f3afc0fa22cb7aff0d00522e571e9b29
{ "arc:challenge": 51.5, "hellaswag": 78.6, "hendrycksTest": 43.6, "truthfulqa:mc": 43.7 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
54.4
NobodyExistsOnTheInternet/PuffedLIMA13bQLORA
main
7da6d235d625e16c850ccd0b947dee40071b1f89
{ "arc:challenge": 59.9, "hellaswag": 84.4, "hendrycksTest": 53.7, "truthfulqa:mc": 39.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59.5
NobodyExistsOnTheInternet/PuffedConvo13bLoraE4
main
40e4fce0c25bd23f6011b424748ee2b5374b98d5
{ "arc:challenge": 59.6, "hellaswag": 84.4, "hendrycksTest": 53.7, "truthfulqa:mc": 39.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59.4
AlpinDale/pygmalion-instruct
main
1665b271316dfee05b2a8daf8b9d6c22ed0aef60
{ "arc:challenge": 52.6, "hellaswag": 77.7, "hendrycksTest": 35.9, "truthfulqa:mc": 42.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
52.1
Doctor-Shotgun/mythospice-70b
main
b00992c26604c9cd496bc41472a05e4c01cd2008
{ "arc:challenge": 69.3, "hellaswag": 87.5, "hendrycksTest": 70.1, "truthfulqa:mc": 56.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
70.9
Doctor-Shotgun/mythospice-limarp-70b
main
ff29fed2a33fc050fd20d0e25b5b23c4a101b074
{ "arc:challenge": 69.2, "hellaswag": 87.5, "hendrycksTest": 70.1, "truthfulqa:mc": 55.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
70.7
voidful/changpt-bart
main
e3d26f736b8b47d5275421be6133b81bef84db7d
{ "arc:challenge": 29.4, "hellaswag": 26.3, "hendrycksTest": 23.1, "truthfulqa:mc": 47.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
31.7
The-Face-Of-Goonery/Huginn-22b-Prototype
main
29222b05794abb862ad0aaaf3020696c9f599810
{ "arc:challenge": 57.7, "hellaswag": 80.7, "hendrycksTest": 49.8, "truthfulqa:mc": 52.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
60.1
The-Face-Of-Goonery/Huginn-v3-13b
main
6c2faf828c5380d28c51fcb4d3d0f1a420fb9a9a
{ "arc:challenge": 60.7, "hellaswag": 82.3, "hendrycksTest": 52.3, "truthfulqa:mc": 50.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61.5
The-Face-Of-Goonery/Huginn-13b-v4.5
main
f3be56d8bf71a8d3905974b1e5fcba7336b02159
{ "arc:challenge": 60.7, "hellaswag": 82.3, "hendrycksTest": 52.3, "truthfulqa:mc": 50.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61.5
The-Face-Of-Goonery/Huginn-13b-v1.2
main
cb3562e7aae05a95fe61610b7b8f4957d3529ce7
{ "arc:challenge": 60.9, "hellaswag": 83.6, "hendrycksTest": 55.3, "truthfulqa:mc": 52 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
63
The-Face-Of-Goonery/huginnv1.2
main
aed4ddc951c657993939fa5b87a4088550569a3b
{ "arc:challenge": 62.4, "hellaswag": 84.3, "hendrycksTest": 57, "truthfulqa:mc": 47.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
62.9
simsim314/WizardLM-70B-V1.0-HF
main
97112db6d0fae8354c13437a5e7dc99fb37b8c2e
{ "arc:challenge": 65.4, "hellaswag": 84.4, "hendrycksTest": 64, "truthfulqa:mc": 54.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
67.2
HyperbeeAI/Tulpar-7b-v0
main
d7c2bc52a3ae13571357f51273ae948caf84400e
{ "arc:challenge": 56.3, "hellaswag": 79, "hendrycksTest": 52.6, "truthfulqa:mc": 51.7 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
59.9
illuin/test-custom-llama
main
d985610bef080473e40f01c53266083c5f0c3169
{ "arc:challenge": 52.3, "hellaswag": 77.5, "hendrycksTest": 36.6, "truthfulqa:mc": 33.8 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
50
luffycodes/mcq-vicuna-13b-v1.5
main
f769a92cfeffe8ee07beee8814ce7eca7cd62805
{ "arc:challenge": 56.2, "hellaswag": 81.1, "hendrycksTest": 53.4, "truthfulqa:mc": 44.1 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
58.7
luffycodes/llama-shishya-7b-ep3-v2
main
679c6cb9e869df686b1ae415ed440e6cfc05f80b
{ "arc:challenge": 47.4, "hellaswag": 75.9, "hendrycksTest": 43.8, "truthfulqa:mc": 30.2 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
49.3
luffycodes/nash-vicuna-13b-v1dot5-ep2-w-rag-w-simple
main
848ef91ab46a72260542283918a971347c6bfa93
{ "arc:challenge": 59.1, "hellaswag": 80.6, "hendrycksTest": 56.1, "truthfulqa:mc": 51.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61.8
luffycodes/llama-shishya-7b-ep3-v1
main
8dc109f45ef36cc7bbd0f5d83fb65ac8e768d1bd
{ "arc:challenge": 48, "hellaswag": 76.6, "hendrycksTest": 46.1, "truthfulqa:mc": 30.9 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
50.4
luffycodes/vicuna-shishya-7b-ep3-v1
main
082cf758aa3f6d8f956056003b5b3b6cde447d88
{ "arc:challenge": 45.9, "hellaswag": 76.4, "hendrycksTest": 50, "truthfulqa:mc": 40.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
53.2
RoversX/llama-2-7b-hf-small-shards-Samantha-V1-SFT
main
c39cee3821269e7fdffa690c2d0836c74dfebd25
{ "arc:challenge": 53.2, "hellaswag": 77.7, "hendrycksTest": 43.5, "truthfulqa:mc": 45.3 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
54.9
circulus/Llama-2-13b-orca-v1
main
e77ec90f432bdffa210a0e4310d117e5d1c662df
{ "arc:challenge": 62.2, "hellaswag": 82.3, "hendrycksTest": 57.7, "truthfulqa:mc": 49.6 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
63
mosaicml/mpt-30b-chat
main
b7957743f18845ff8695f7919420adb769ec225e
{ "arc:challenge": 58.4, "hellaswag": 82.4, "hendrycksTest": 51, "truthfulqa:mc": 52 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
61
mosaicml/mpt-30b
main
0261af71d7177453889f868d26607dec8d5aaa2e
{ "arc:challenge": 55.9, "hellaswag": 82.4, "hendrycksTest": 47.9, "truthfulqa:mc": 38.4 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
56.2
mosaicml/mpt-7b-8k-chat
main
ef97b878a279cd1765fbed7b8321fb3cff1aa5b5
{ "arc:challenge": 48, "hellaswag": 77.6, "hendrycksTest": 41.9, "truthfulqa:mc": 43.7 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
52.8
mosaicml/mpt-7b-chat
main
64e5c9c9fb53a8e89690c2dee75a5add37f7113e
{ "arc:challenge": 46.5, "hellaswag": 75.5, "hendrycksTest": 37.6, "truthfulqa:mc": 40.2 }
9ba100d35ce48d3d4c132947464c93c861932caa
2023-11-23T17:28:23
50