model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
h2oai/h2ogpt-research-oasst1-llama-65b | main | a6d8676aaa2ca2c25ea99180b538f0369dc70185 | {
"arc:challenge": 64.8,
"hellaswag": 85.9,
"hendrycksTest": 63.6,
"truthfulqa:mc": 48.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 65.8 |
h2oai/h2ogpt-oig-oasst1-256-6_9b | main | f1c9bac89b74d3487cb092788ce828fb9520c1a7 | {
"arc:challenge": 39.9,
"hellaswag": 65.4,
"hendrycksTest": 26.4,
"truthfulqa:mc": 35
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 41.7 |
h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt | main | 754e0c90ed5d9241fdfd5a188572b3ea2152eaa7 | {
"arc:challenge": 34,
"hellaswag": 50.5,
"hendrycksTest": 24.7,
"truthfulqa:mc": 41.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 37.8 |
h2oai/h2ogpt-research-oig-oasst1-512-30b | main | 4215e83b9038c9e61d979cf5223b29f860bace9b | {
"arc:challenge": 59,
"hellaswag": 82.6,
"hendrycksTest": 50.7,
"truthfulqa:mc": 48.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 60.2 |
h2oai/h2ogpt-gm-oasst1-en-1024-20b | main | 1a5b8d25587eab67d837621a6c9423e7ef6df289 | {
"arc:challenge": 48,
"hellaswag": 72.8,
"hendrycksTest": 26,
"truthfulqa:mc": 39.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 46.7 |
h2oai/h2ogpt-oasst1-512-12b | main | c6bb0fe363e0105839d34ca757793b61c9606f95 | {
"arc:challenge": 42.3,
"hellaswag": 70.2,
"hendrycksTest": 26,
"truthfulqa:mc": 36.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 43.7 |
h2oai/h2ogpt-gm-oasst1-en-1024-12b | main | e547fffafb382fd39ef5de35ba3b5afc1b43e74d | {
"arc:challenge": 43.1,
"hellaswag": 69.7,
"hendrycksTest": 25.9,
"truthfulqa:mc": 38
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 44.2 |
minlik/chinese-alpaca-33b-merged | main | fc2535104c0b48afc42575f9fe10bbcbb7612ec3 | {
"arc:challenge": 59.3,
"hellaswag": 78.4,
"hendrycksTest": 57.7,
"truthfulqa:mc": 52.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 62 |
xzuyn/Alpacino-SuperCOT-13B | main | 3a82b04684fe99d59556421c3f96a187049a3cec | {
"arc:challenge": 58.4,
"hellaswag": 81.7,
"hendrycksTest": 47.9,
"truthfulqa:mc": 45.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.4 |
xzuyn/MedicWizard-7B | main | 0b3ef975fb5e8ac1eae775160ab54c98221889df | {
"arc:challenge": 53.5,
"hellaswag": 78.4,
"hendrycksTest": 44.6,
"truthfulqa:mc": 41.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.4 |
teknium/OpenHermes-2.5-Mistral-7B | main | 2a54cad766bc90828354db5c4199795aecfd0df1 | {
"arc:challenge": 64.9,
"hellaswag": 84.3,
"hendrycksTest": 63.8,
"truthfulqa:mc": 52.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 66.3 |
teknium/OpenHermes-2-Mistral-7B | main | 843a9bb94fac7d7bfc1b7c9f201efba295b6f5d6 | {
"arc:challenge": 63.1,
"hellaswag": 83.8,
"hendrycksTest": 63.5,
"truthfulqa:mc": 50.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 65.2 |
teknium/OpenHermes-13B | main | f09d0fe655ad57cce9179b7b40ea6f81e07db18c | {
"arc:challenge": 59.8,
"hellaswag": 82.2,
"hendrycksTest": 56.3,
"truthfulqa:mc": 46
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.1 |
teknium/CollectiveCognition-v1.1-Mistral-7B | main | 5f57f70ec99450c70da2540e94dd7fd67be4b23c | {
"arc:challenge": 62.1,
"hellaswag": 84.2,
"hendrycksTest": 62.3,
"truthfulqa:mc": 57.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 66.6 |
nathan0/mpt_delta_tuned_model_v2 | main | 6adb4cb4ba37f4ce9e9c3051d343addf1098182c | {
"arc:challenge": 50.7,
"hellaswag": 76.4,
"hendrycksTest": 28.7,
"truthfulqa:mc": 35.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 47.8 |
nathan0/mpt_delta_tuned_model_v3 | main | 6adb4cb4ba37f4ce9e9c3051d343addf1098182c | {
"arc:challenge": 50.7,
"hellaswag": 76.4,
"hendrycksTest": 28.7,
"truthfulqa:mc": 35.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 47.8 |
chargoddard/duplicitous-mammal-13b | main | a05d0562b8da2ac2e76aa65984e8063249bc85c8 | {
"arc:challenge": 61.7,
"hellaswag": 83.8,
"hendrycksTest": 57.5,
"truthfulqa:mc": 52.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.8 |
chargoddard/llama2-22b-blocktriangular | main | 40a51343ae776b5cb39f2b4343ae8f9b676ffd58 | {
"arc:challenge": 58.5,
"hellaswag": 82.6,
"hendrycksTest": 54.6,
"truthfulqa:mc": 39.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.8 |
chargoddard/MelangeB-70b | main | 08239fb1e30b1e42b14370f23e942bc51e76027c | {
"arc:challenge": 71.7,
"hellaswag": 87.5,
"hendrycksTest": 70,
"truthfulqa:mc": 59.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 72.1 |
chargoddard/ypotryll-22b-epoch2-qlora | main | 26fdd8fa420d72ed835c7d17086f0441db0985d4 | {
"arc:challenge": 59.2,
"hellaswag": 80.7,
"hendrycksTest": 54.5,
"truthfulqa:mc": 40.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.7 |
chargoddard/llama-2-26b-trenchcoat-stack | main | 075d67c3223f4b379ab7f997c3787cd0630d80f7 | {
"arc:challenge": 55,
"hellaswag": 79.9,
"hendrycksTest": 53.7,
"truthfulqa:mc": 40.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 57.3 |
chargoddard/llama2-22b | main | 2bece0787009b4b584f49d0e0d1b49ecf4a52da9 | {
"arc:challenge": 58.5,
"hellaswag": 82.6,
"hendrycksTest": 54.7,
"truthfulqa:mc": 39.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.9 |
chargoddard/platypus-2-22b-relora | main | 15bca3e9b25cc2f280fec21686ef3bc445217503 | {
"arc:challenge": 57.7,
"hellaswag": 82.4,
"hendrycksTest": 55.3,
"truthfulqa:mc": 43.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.8 |
chargoddard/platypus2-22b-relora | main | 15bca3e9b25cc2f280fec21686ef3bc445217503 | {
"arc:challenge": 57.5,
"hellaswag": 82.4,
"hendrycksTest": 54.9,
"truthfulqa:mc": 43.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.6 |
chargoddard/storytime-13b | main | 233568319a636b6a7b02a4def2c51d08a3e0fbfc | {
"arc:challenge": 62,
"hellaswag": 84,
"hendrycksTest": 57.5,
"truthfulqa:mc": 52.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64 |
chargoddard/Chronorctypus-Limarobormes-13b | main | 75c1bf5f4b40cf61873ff6487ccd3efc4f684330 | {
"arc:challenge": 59.9,
"hellaswag": 82.8,
"hendrycksTest": 58.4,
"truthfulqa:mc": 51.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.2 |
chargoddard/MelangeC-70b | main | e54a2b924dec135f3fa2373933ab8485178cde1b | {
"arc:challenge": 71.7,
"hellaswag": 87.6,
"hendrycksTest": 70.4,
"truthfulqa:mc": 58.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 72 |
Open-Orca/Mistral-7B-SlimOrca | main | a9744d8cf9ce4230678a891bcf8bba7cbc0aaece | {
"arc:challenge": 62.5,
"hellaswag": 83.9,
"hendrycksTest": 62.8,
"truthfulqa:mc": 54.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 65.8 |
Open-Orca/LlongOrca-13B-16k | main | 8ea1fb205553cadbc90069d80a7e58281b6281c3 | {
"arc:challenge": 62.5,
"hellaswag": 82.8,
"hendrycksTest": 55.5,
"truthfulqa:mc": 50.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 62.7 |
Open-Orca/OpenOrca-Preview1-13B | main | 4c558283a98348383460939afda9cb5c54544c8f | {
"arc:challenge": 54.9,
"hellaswag": 78.2,
"hendrycksTest": 50.1,
"truthfulqa:mc": 49.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.1 |
Open-Orca/OpenOrca-Platypus2-13B | main | e7a40134f7eb687c6ab66d445dc7251257f8d391 | {
"arc:challenge": 62.8,
"hellaswag": 83.2,
"hendrycksTest": 59.4,
"truthfulqa:mc": 53.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.6 |
Open-Orca/LlongOrca-7B-16k | main | 1370c7c595e6c8394e6332bc535ae25e21def85b | {
"arc:challenge": 57.5,
"hellaswag": 79.4,
"hendrycksTest": 49.4,
"truthfulqa:mc": 49.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59 |
grantprice/Cerebras-GPT-590M-finetuned-DND | main | a0a2fbe342cdc86433913ba5f96978e4703ff672 | {
"arc:challenge": 24.7,
"hellaswag": 27.8,
"hendrycksTest": 23.1,
"truthfulqa:mc": 48.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 31 |
lgaalves/tinyllama-1.1b-chat-v0.3_platypus | main | 0bb6ebe1d41d394bae0ed9107ec8d776d9d76a68 | {
"arc:challenge": 30.3,
"hellaswag": 55.1,
"hendrycksTest": 26.1,
"truthfulqa:mc": 39.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 37.7 |
lgaalves/gpt2_guanaco-dolly-platypus | main | 6bf0a8146cf255c829ec2ad83926c8b80945b431 | {
"arc:challenge": 23.5,
"hellaswag": 31,
"hendrycksTest": 26.4,
"truthfulqa:mc": 40
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.2 |
lgaalves/mistral-7b_open_platypus | main | b9a60b9ad0fe06bd314ffe99d543f1df6ecd10da | {
"arc:challenge": 55.8,
"hellaswag": 82.1,
"hendrycksTest": 59.8,
"truthfulqa:mc": 48.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.6 |
lgaalves/gpt2_camel_physics-platypus | main | 66165ff32ed8de6c39f3524a810f5e97ba6d3347 | {
"arc:challenge": 23,
"hellaswag": 31.3,
"hendrycksTest": 26.9,
"truthfulqa:mc": 39.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.2 |
lgaalves/gpt2-xl_lima | main | f7db5b1db521abd7578b95138e737637e0037ca5 | {
"arc:challenge": 31.1,
"hellaswag": 51.3,
"hendrycksTest": 25.4,
"truthfulqa:mc": 38.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 36.6 |
lgaalves/llama-2-7b-hf_open-platypus | main | c7e776f3f3afc0fa22cb7aff0d00522e571e9b29 | {
"arc:challenge": 51.5,
"hellaswag": 78.6,
"hendrycksTest": 43.6,
"truthfulqa:mc": 43.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.4 |
NobodyExistsOnTheInternet/PuffedLIMA13bQLORA | main | 7da6d235d625e16c850ccd0b947dee40071b1f89 | {
"arc:challenge": 59.9,
"hellaswag": 84.4,
"hendrycksTest": 53.7,
"truthfulqa:mc": 39.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.5 |
NobodyExistsOnTheInternet/PuffedConvo13bLoraE4 | main | 40e4fce0c25bd23f6011b424748ee2b5374b98d5 | {
"arc:challenge": 59.6,
"hellaswag": 84.4,
"hendrycksTest": 53.7,
"truthfulqa:mc": 39.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.4 |
AlpinDale/pygmalion-instruct | main | 1665b271316dfee05b2a8daf8b9d6c22ed0aef60 | {
"arc:challenge": 52.6,
"hellaswag": 77.7,
"hendrycksTest": 35.9,
"truthfulqa:mc": 42.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 52.1 |
Doctor-Shotgun/mythospice-70b | main | b00992c26604c9cd496bc41472a05e4c01cd2008 | {
"arc:challenge": 69.3,
"hellaswag": 87.5,
"hendrycksTest": 70.1,
"truthfulqa:mc": 56.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 70.9 |
Doctor-Shotgun/mythospice-limarp-70b | main | ff29fed2a33fc050fd20d0e25b5b23c4a101b074 | {
"arc:challenge": 69.2,
"hellaswag": 87.5,
"hendrycksTest": 70.1,
"truthfulqa:mc": 55.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 70.7 |
voidful/changpt-bart | main | e3d26f736b8b47d5275421be6133b81bef84db7d | {
"arc:challenge": 29.4,
"hellaswag": 26.3,
"hendrycksTest": 23.1,
"truthfulqa:mc": 47.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 31.7 |
The-Face-Of-Goonery/Huginn-22b-Prototype | main | 29222b05794abb862ad0aaaf3020696c9f599810 | {
"arc:challenge": 57.7,
"hellaswag": 80.7,
"hendrycksTest": 49.8,
"truthfulqa:mc": 52.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 60.1 |
The-Face-Of-Goonery/Huginn-v3-13b | main | 6c2faf828c5380d28c51fcb4d3d0f1a420fb9a9a | {
"arc:challenge": 60.7,
"hellaswag": 82.3,
"hendrycksTest": 52.3,
"truthfulqa:mc": 50.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.5 |
The-Face-Of-Goonery/Huginn-13b-v4.5 | main | f3be56d8bf71a8d3905974b1e5fcba7336b02159 | {
"arc:challenge": 60.7,
"hellaswag": 82.3,
"hendrycksTest": 52.3,
"truthfulqa:mc": 50.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.5 |
The-Face-Of-Goonery/Huginn-13b-v1.2 | main | cb3562e7aae05a95fe61610b7b8f4957d3529ce7 | {
"arc:challenge": 60.9,
"hellaswag": 83.6,
"hendrycksTest": 55.3,
"truthfulqa:mc": 52
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63 |
The-Face-Of-Goonery/huginnv1.2 | main | aed4ddc951c657993939fa5b87a4088550569a3b | {
"arc:challenge": 62.4,
"hellaswag": 84.3,
"hendrycksTest": 57,
"truthfulqa:mc": 47.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 62.9 |
simsim314/WizardLM-70B-V1.0-HF | main | 97112db6d0fae8354c13437a5e7dc99fb37b8c2e | {
"arc:challenge": 65.4,
"hellaswag": 84.4,
"hendrycksTest": 64,
"truthfulqa:mc": 54.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 67.2 |
HyperbeeAI/Tulpar-7b-v0 | main | d7c2bc52a3ae13571357f51273ae948caf84400e | {
"arc:challenge": 56.3,
"hellaswag": 79,
"hendrycksTest": 52.6,
"truthfulqa:mc": 51.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.9 |
illuin/test-custom-llama | main | d985610bef080473e40f01c53266083c5f0c3169 | {
"arc:challenge": 52.3,
"hellaswag": 77.5,
"hendrycksTest": 36.6,
"truthfulqa:mc": 33.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 50 |
luffycodes/mcq-vicuna-13b-v1.5 | main | f769a92cfeffe8ee07beee8814ce7eca7cd62805 | {
"arc:challenge": 56.2,
"hellaswag": 81.1,
"hendrycksTest": 53.4,
"truthfulqa:mc": 44.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.7 |
luffycodes/llama-shishya-7b-ep3-v2 | main | 679c6cb9e869df686b1ae415ed440e6cfc05f80b | {
"arc:challenge": 47.4,
"hellaswag": 75.9,
"hendrycksTest": 43.8,
"truthfulqa:mc": 30.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 49.3 |
luffycodes/nash-vicuna-13b-v1dot5-ep2-w-rag-w-simple | main | 848ef91ab46a72260542283918a971347c6bfa93 | {
"arc:challenge": 59.1,
"hellaswag": 80.6,
"hendrycksTest": 56.1,
"truthfulqa:mc": 51.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.8 |
luffycodes/llama-shishya-7b-ep3-v1 | main | 8dc109f45ef36cc7bbd0f5d83fb65ac8e768d1bd | {
"arc:challenge": 48,
"hellaswag": 76.6,
"hendrycksTest": 46.1,
"truthfulqa:mc": 30.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 50.4 |
luffycodes/vicuna-shishya-7b-ep3-v1 | main | 082cf758aa3f6d8f956056003b5b3b6cde447d88 | {
"arc:challenge": 45.9,
"hellaswag": 76.4,
"hendrycksTest": 50,
"truthfulqa:mc": 40.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 53.2 |
RoversX/llama-2-7b-hf-small-shards-Samantha-V1-SFT | main | c39cee3821269e7fdffa690c2d0836c74dfebd25 | {
"arc:challenge": 53.2,
"hellaswag": 77.7,
"hendrycksTest": 43.5,
"truthfulqa:mc": 45.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.9 |
circulus/Llama-2-13b-orca-v1 | main | e77ec90f432bdffa210a0e4310d117e5d1c662df | {
"arc:challenge": 62.2,
"hellaswag": 82.3,
"hendrycksTest": 57.7,
"truthfulqa:mc": 49.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63 |
mosaicml/mpt-30b-chat | main | b7957743f18845ff8695f7919420adb769ec225e | {
"arc:challenge": 58.4,
"hellaswag": 82.4,
"hendrycksTest": 51,
"truthfulqa:mc": 52
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61 |
mosaicml/mpt-30b | main | 0261af71d7177453889f868d26607dec8d5aaa2e | {
"arc:challenge": 55.9,
"hellaswag": 82.4,
"hendrycksTest": 47.9,
"truthfulqa:mc": 38.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 56.2 |
mosaicml/mpt-7b-8k-chat | main | ef97b878a279cd1765fbed7b8321fb3cff1aa5b5 | {
"arc:challenge": 48,
"hellaswag": 77.6,
"hendrycksTest": 41.9,
"truthfulqa:mc": 43.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 52.8 |
mosaicml/mpt-7b-chat | main | 64e5c9c9fb53a8e89690c2dee75a5add37f7113e | {
"arc:challenge": 46.5,
"hellaswag": 75.5,
"hendrycksTest": 37.6,
"truthfulqa:mc": 40.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 50 |