model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
RWKV/rwkv-4-169m-pile | main | 46bdc280eb97b6141d5d51a935e0c4870ecaefcc | {
"arc:challenge": 23.6,
"hellaswag": 31.7,
"hendrycksTest": 23.2,
"truthfulqa:mc": 41.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.1 |
RWKV/rwkv-4-430m-pile | main | a4f6ec80438d4262d1bbc8f385feb2ef1a4a9d6b | {
"arc:challenge": 26.7,
"hellaswag": 40,
"hendrycksTest": 24.9,
"truthfulqa:mc": 39.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 32.8 |
RWKV/rwkv-4-7b-pile | main | 922e22a761427e50d7be457b31a76b1126021b8b | {
"arc:challenge": 39.7,
"hellaswag": 66.3,
"hendrycksTest": 25,
"truthfulqa:mc": 33.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 41.2 |
RWKV/rwkv-raven-14b | main | 359c0649b4f1d10a26ebea32908035bc00d152ee | {
"arc:challenge": 44.6,
"hellaswag": 71.3,
"hendrycksTest": 25.9,
"truthfulqa:mc": 41.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 45.9 |
RWKV/rwkv-raven-1b5 | main | 571a3bd891ce33f2ee3fc6de09218178edb0dae2 | {
"arc:challenge": 31.8,
"hellaswag": 52.6,
"hendrycksTest": 26,
"truthfulqa:mc": 37.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 36.9 |
jaspercatapang/Echidna-30B | main | 20b13b6676d54b555ae2b9b2b4b6fc8a0c7c2e89 | {
"arc:challenge": 28.5,
"hellaswag": 25.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 31.8 |
CoolWP/llama-2-13b-guanaco-fp16 | main | a60e8e39e4fbe271655e1c78eb1ceb2081518231 | {
"arc:challenge": 59.6,
"hellaswag": 82.4,
"hendrycksTest": 55.5,
"truthfulqa:mc": 43.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 60.2 |
TigerResearch/tigerbot-7b-sft | main | 98b847905d63f74624e834db1ff95ee2814cbbd3 | {
"arc:challenge": 41.6,
"hellaswag": 60.6,
"hendrycksTest": 29.9,
"truthfulqa:mc": 58.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 47.6 |
TigerResearch/tigerbot-7b-base | main | 300831494aa1eb16e59799310a09531f60dcc904 | {
"arc:challenge": 47.7,
"hellaswag": 72.1,
"hendrycksTest": 45.1,
"truthfulqa:mc": 42.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 51.8 |
MayaPH/FinOPT-Washington | main | cdd8a6cde7902de39757cf31d73af1f51df0d8e8 | {
"arc:challenge": 25.2,
"hellaswag": 26.2,
"hendrycksTest": 24.8,
"truthfulqa:mc": 45.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.5 |
MayaPH/GodziLLa-30B | main | aa9912a2ac60abeac28b4566731cd903dcc582ac | {
"arc:challenge": 61.5,
"hellaswag": 82.1,
"hendrycksTest": 54.2,
"truthfulqa:mc": 55.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.4 |
MayaPH/GodziLLa-30B-instruct | main | 642bf3683801e20e4b7cf28d94374d5e6054c007 | {
"arc:challenge": 29,
"hellaswag": 26.5,
"hendrycksTest": 24.9,
"truthfulqa:mc": 48.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 32.3 |
MayaPH/GodziLLa-30B-plus | main | a66b1860d11ebf8aed07237cf636fdd2b3a07f06 | {
"arc:challenge": 28.9,
"hellaswag": 26.4,
"hendrycksTest": 24.6,
"truthfulqa:mc": 48.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 32.2 |
MayaPH/opt-flan-iml-6.7b | main | cbe8d60db6f3c52e653ca73e23a1c34c08127d02 | {
"arc:challenge": 30.1,
"hellaswag": 58.8,
"hendrycksTest": 25.1,
"truthfulqa:mc": 36.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 37.7 |
MayaPH/GodziLLa2-70B | main | 7b78087db07eec97f7b461d10758ece76d685543 | {
"arc:challenge": 71.4,
"hellaswag": 87.5,
"hendrycksTest": 69.9,
"truthfulqa:mc": 61.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 72.6 |
abhishek/llama2guanacotest | main | 679d17809939a0bf9b79bbb027898cbea64045b2 | {
"arc:challenge": 51.6,
"hellaswag": 77.6,
"hendrycksTest": 48.5,
"truthfulqa:mc": 43.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 55.4 |
abhishek/autotrain-llama-alpaca-peft-52508123785 | main | {
"arc:challenge": 52.2,
"hellaswag": 76.9,
"hendrycksTest": 37.6,
"truthfulqa:mc": 32.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 49.9 |
|
guardrail/llama-2-7b-guanaco-instruct-sharded | main | fc7a3abbc3b9a9b3e163ef3c4844307ac270fca7 | {
"arc:challenge": 53.8,
"hellaswag": 78.7,
"hendrycksTest": 46.7,
"truthfulqa:mc": 43.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 55.8 |
CobraMamba/mamba-gpt-3b-v2 | main | 935f4d90bd0fc7117113d3c7b6b6af9dba93183d | {
"arc:challenge": 42.2,
"hellaswag": 71.5,
"hendrycksTest": 27.1,
"truthfulqa:mc": 36.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 44.4 |
CobraMamba/mamba-gpt-7b | main | cb0b04b1bff7921614efbd87d5b87bac04c58d13 | {
"arc:challenge": 51.2,
"hellaswag": 75.4,
"hendrycksTest": 47.5,
"truthfulqa:mc": 42.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54 |
CobraMamba/mamba-gpt-7b-v1 | main | e64d658b397748e409d9633fd24fc5a6df429600 | {
"arc:challenge": 61.3,
"hellaswag": 84.1,
"hendrycksTest": 63.5,
"truthfulqa:mc": 46.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.8 |
CobraMamba/mamba-gpt-3b-v3 | main | d860a90ef6b30c695b985dd2ff382d4bbb80e857 | {
"arc:challenge": 41.7,
"hellaswag": 71.1,
"hendrycksTest": 27.3,
"truthfulqa:mc": 37.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 44.5 |
CobraMamba/mamba-gpt-7b-v2 | main | 6439444e2c0b61253d3e61ae04fe0436717acc2f | {
"arc:challenge": 61.9,
"hellaswag": 83.8,
"hendrycksTest": 61.7,
"truthfulqa:mc": 46.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.5 |
blueapple8259/TinyStories-Alpaca | main | 18e0bde7e72e477757832f0624a0410efc066216 | {
"arc:challenge": 24,
"hellaswag": 24.9,
"hendrycksTest": 23.4,
"truthfulqa:mc": 46.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 29.8 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-FT-LoRA-8bit-test1 | main | f477d24b00e05fe4c5f8d5f933080994cfd90e4e | {
"arc:challenge": 38.7,
"hellaswag": 63.5,
"hendrycksTest": 25.2,
"truthfulqa:mc": 36.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 40.9 |
DanielSc4/RedPajama-INCITE-Chat-3B-v1-RL-LoRA-8bit-test1 | main | a2ee88a9fa1c9ad41e0a8c15217a4b1230ec33c8 | {
"arc:challenge": 41.3,
"hellaswag": 66.8,
"hendrycksTest": 26.1,
"truthfulqa:mc": 35
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 42.3 |
THUDM/chatglm2-6b | main | 162b620e3078b03eefff94eb5f762d4093425fb5 | {
"arc:challenge": 38.8,
"hellaswag": 59,
"hendrycksTest": 46.7,
"truthfulqa:mc": 48.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 48.2 |
Salesforce/codegen-6B-multi | main | 2d58b1e73791e8f0be7ea59c2720dccb6f4d0f06 | {
"arc:challenge": 27.2,
"hellaswag": 41.1,
"hendrycksTest": 25.7,
"truthfulqa:mc": 45.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 34.9 |
mrm8488/llama-2-coder-7b | main | f21c0d5e3f9f8c5addf093358e6885afa9602296 | {
"arc:challenge": 54,
"hellaswag": 78.4,
"hendrycksTest": 46.3,
"truthfulqa:mc": 38.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.3 |
mrm8488/mistral-7b-ft-h4-no_robots_instructions | main | 785446da9a53ceae48795069bf7ccaf46a91a5ba | {
"arc:challenge": 60.9,
"hellaswag": 83.2,
"hendrycksTest": 63.7,
"truthfulqa:mc": 43.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 62.8 |
Riiid/sheep-duck-llama-2 | main | e196dd0fe1d604c4975d972b177b09e4f1572cd5 | {
"arc:challenge": 72.4,
"hellaswag": 87.8,
"hendrycksTest": 70.8,
"truthfulqa:mc": 63.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 73.7 |
Riiid/sheep-duck-llama-2-70b-v1.1 | main | a8fc5c02c995733af6339ec882bef4ed93db1e8f | {
"arc:challenge": 73,
"hellaswag": 87.8,
"hendrycksTest": 70.8,
"truthfulqa:mc": 64.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 74.1 |
PSanni/Deer-3b | main | 53ea8f8862fc1820f0cd31f62953b7290fd79867 | {
"arc:challenge": 38.5,
"hellaswag": 57.4,
"hendrycksTest": 25.6,
"truthfulqa:mc": 40
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 40.4 |
w601sxs/b1ade-1b | main | b4b0fd71589e6590089e1ec14a840ecab10894ae | {
"arc:challenge": 28.6,
"hellaswag": 46.1,
"hendrycksTest": 25.1,
"truthfulqa:mc": 41.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 35.3 |
bofenghuang/vigogne-13b-chat | main | 27002e974774c3599e6a4d731dd44e68b9e41f92 | {
"arc:challenge": 58.6,
"hellaswag": 80.9,
"hendrycksTest": 47.8,
"truthfulqa:mc": 48.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59 |
bofenghuang/vigogne-33b-instruct | main | 9c2b558b888e0ef8b4a72e0771db72a06a5c8474 | {
"arc:challenge": 63.1,
"hellaswag": 85,
"hendrycksTest": 58.3,
"truthfulqa:mc": 52.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.6 |
bofenghuang/vigostral-7b-chat | main | 969fbfc7a91f53c8562a2c48a3c24dd3745d5a97 | {
"arc:challenge": 62.6,
"hellaswag": 84.3,
"hendrycksTest": 63.5,
"truthfulqa:mc": 49.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.9 |
u-chom/preded-title-amazongoogle-abtbuy | main | ab36321d76775d6e276d157e27de23854d21be3a | {
"arc:challenge": 50.9,
"hellaswag": 78.1,
"hendrycksTest": 38,
"truthfulqa:mc": 41.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 52.2 |
yeen214/test_llama2_7b | main | 69a4886f51ed752216cdd7f41a584d14240126f9 | {
"arc:challenge": 53.1,
"hellaswag": 78.6,
"hendrycksTest": 46.9,
"truthfulqa:mc": 38.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.4 |
yeen214/llama2_7b_merge_orcafamily | main | fb65f697de632f2f3fef57fc3cd12fb5e4913a89 | {
"arc:challenge": 56.9,
"hellaswag": 81.2,
"hendrycksTest": 51.5,
"truthfulqa:mc": 49.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.8 |
DevaMalla/llama_7b_qlora_cds | main | b6b5c65c5c1cce34d24c8f790bb0cc011e0f0808 | {
"arc:challenge": 52.5,
"hellaswag": 77.8,
"hendrycksTest": 32.4,
"truthfulqa:mc": 46.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 52.2 |
DevaMalla/llama_7b_qlora_pds-eval | main | d20419e1d9e9a6a59ced3edf5169e8e7b3e8394c | {
"arc:challenge": 53.9,
"hellaswag": 78.1,
"hendrycksTest": 33,
"truthfulqa:mc": 45.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 52.6 |
DevaMalla/llama-base-7b | main | e01d89d8e444f7d751ea58feaf22ff8c9af69d2a | {
"arc:challenge": 50.9,
"hellaswag": 77.8,
"hendrycksTest": 35.7,
"truthfulqa:mc": 34.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 49.7 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-QLoRA-multigpu | main | f65029ea8f030731ace568e40bab33a7097a13de | {
"arc:challenge": 57.5,
"hellaswag": 82.5,
"hendrycksTest": 54.8,
"truthfulqa:mc": 43.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.6 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus | main | c318a24121bd69509f395e17a9636093213ece21 | {
"arc:challenge": 58.9,
"hellaswag": 82.1,
"hendrycksTest": 55,
"truthfulqa:mc": 42.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.7 |
NekoPunchBBB/Llama-2-13b-hf_Open-Platypus-8bit-att | main | 83a8e51d0a72dcfbe5de13dc7ee10dc20e91602e | {
"arc:challenge": 57.5,
"hellaswag": 82.1,
"hendrycksTest": 54.6,
"truthfulqa:mc": 42.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.1 |
NewstaR/Koss-7B-chat | main | b1ab836d9ebf7029fafa07949b51d3838501d537 | {
"arc:challenge": 53.7,
"hellaswag": 78.8,
"hendrycksTest": 46.7,
"truthfulqa:mc": 44
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 55.8 |
NewstaR/Starlight-13B | main | cb9fced568b1abd881133c642c427aaa488f00cc | {
"arc:challenge": 59.3,
"hellaswag": 82.2,
"hendrycksTest": 55.7,
"truthfulqa:mc": 37.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.6 |
NewstaR/Morningstar-13b-hf | main | 2605b5b3b0ecba906ac26d39aab40f33c2ec81c9 | {
"arc:challenge": 59,
"hellaswag": 81.9,
"hendrycksTest": 54.6,
"truthfulqa:mc": 44.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.9 |
NewstaR/Starlight-7B | main | 1f7436c458ebc3d8d31b91091c1a7a48e942cd3b | {
"arc:challenge": 53.1,
"hellaswag": 78.6,
"hendrycksTest": 46.8,
"truthfulqa:mc": 38.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 54.3 |
Sao10K/Stheno-1.8-L2-13B | main | fe054ab749a69375285df40913a88bd40f1e2bf6 | {
"arc:challenge": 63.5,
"hellaswag": 84.1,
"hendrycksTest": 58.6,
"truthfulqa:mc": 52.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.8 |
Sao10K/Mythical-Destroyer-V2-L2-13B | main | cbc8b2e4a3beafc311b9e61f8fa9f7526a77c360 | {
"arc:challenge": 59.3,
"hellaswag": 82.7,
"hendrycksTest": 57.4,
"truthfulqa:mc": 57.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.1 |
Sao10K/Stheno-Inverted-1.2-L2-13B | main | 8d2e9087093eef1c9173e167beb40b9d034a4655 | {
"arc:challenge": 59.4,
"hellaswag": 83,
"hendrycksTest": 55.8,
"truthfulqa:mc": 51.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 62.4 |
Sao10K/Zephyrus-L1-33B | main | 679aae34440d576456b283070371b2a15dbb948b | {
"arc:challenge": 64.5,
"hellaswag": 84.1,
"hendrycksTest": 57.4,
"truthfulqa:mc": 53.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 65 |
Sao10K/SthenoWriter-L2-13B | main | a6d9e26ab765eb170cc0aa428ee5e25b08524657 | {
"arc:challenge": 62.3,
"hellaswag": 83.3,
"hendrycksTest": 56.1,
"truthfulqa:mc": 44.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.6 |
Sao10K/Chat-Stheno-L2-13B | main | 20419fdd5b4bdcbbf075223c33b396958c48a6cf | {
"arc:challenge": 58.4,
"hellaswag": 81,
"hendrycksTest": 54.8,
"truthfulqa:mc": 43.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.4 |
Sao10K/Euryale-1.3-L2-70B | main | 6e3ce78eb5346bf3a5ee88cd60c25dc0d73de639 | {
"arc:challenge": 70.8,
"hellaswag": 87.9,
"hendrycksTest": 70.4,
"truthfulqa:mc": 59.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 72.2 |
Sao10K/Mythical-Destroyer-L2-13B | main | 7c87376b201b1c30c4e12c0b7bc2f28f017ce7bc | {
"arc:challenge": 58.7,
"hellaswag": 82,
"hendrycksTest": 57.7,
"truthfulqa:mc": 56.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.7 |
Sao10K/BrainDerp2 | main | 948ee7af94a8b092807df4becfc0a8c1cd042878 | {
"arc:challenge": 60.9,
"hellaswag": 81.9,
"hendrycksTest": 58.9,
"truthfulqa:mc": 57.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.7 |
Sao10K/Euryale-L2-70B | main | 6589310a57ce5d9d6877f353f3d00cda8fa9101c | {
"arc:challenge": 68.9,
"hellaswag": 87.1,
"hendrycksTest": 68.8,
"truthfulqa:mc": 54.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 69.8 |
Sao10K/Stheno-Mix-L2-20B | main | 6f9dcdaae6ef9071effe63d2107abe8b9712345b | {
"arc:challenge": 57.8,
"hellaswag": 79.6,
"hendrycksTest": 52.5,
"truthfulqa:mc": 51.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 60.4 |
nomic-ai/gpt4all-j | main | c7244e40ac6f3a52caecc96cd45481caae57ae8c | {
"arc:challenge": 42,
"hellaswag": 64.1,
"hendrycksTest": 28.2,
"truthfulqa:mc": 42.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 44.3 |
tianyil1/denas-llama2 | main | b8aebc9157c0e427536aeac9132021fd66615702 | {
"arc:challenge": 53.9,
"hellaswag": 77.8,
"hendrycksTest": 45.5,
"truthfulqa:mc": 45.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 55.6 |
llm-agents/tora-70b-v1.0 | main | e95fd7daf017e7c414ec07ebef4ddf013c16f9a4 | {
"arc:challenge": 67.7,
"hellaswag": 85.8,
"hendrycksTest": 69.2,
"truthfulqa:mc": 51.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 68.6 |
rishiraj/bloom-560m-guanaco | main | 17b886fe53bdb4cea75a7f40da1e8e987124edef | {
"arc:challenge": 27.9,
"hellaswag": 26.1,
"hendrycksTest": 24.5,
"truthfulqa:mc": 49.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 32 |
ByteWave/Yi-8B-Llama | main | 4f3f4d73ff3962487d1c51702b02d795bf1f33a4 | {
"arc:challenge": 25.7,
"hellaswag": 26.8,
"hendrycksTest": 24.1,
"truthfulqa:mc": 47.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 31.1 |
codellama/CodeLlama-34b-hf | main | c778b02fdecd4663d2b0a42bfb340fd29969533b | {
"arc:challenge": 37.5,
"hellaswag": 31.8,
"hendrycksTest": 37.2,
"truthfulqa:mc": 38.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 36.4 |
codellama/CodeLlama-7b-Instruct-hf | main | 7affc442e639b8aa1c4b3e98a10a2f45a21b8b4f | {
"arc:challenge": 36.5,
"hellaswag": 55.4,
"hendrycksTest": 34.5,
"truthfulqa:mc": 41.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 41.9 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-2.2epochs-oasst1-top1-instruct-V1 | main | 74cd9eba94e77832b3081689fc5c99c37c063790 | {
"arc:challenge": 31.5,
"hellaswag": 54.4,
"hendrycksTest": 25.5,
"truthfulqa:mc": 42.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 38.4 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-3epochs-oasst1-top1-instruct-V1 | main | b1ec2a1e08eb790b9a32a43053316650921af943 | {
"arc:challenge": 31.4,
"hellaswag": 54.2,
"hendrycksTest": 25.4,
"truthfulqa:mc": 42.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 38.4 |
habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-4epochs-oasst1-top1-instruct-V1 | main | 7cd6d5ad10180127771e4326772eae3d40fa8445 | {
"arc:challenge": 31.1,
"hellaswag": 54.3,
"hendrycksTest": 25.4,
"truthfulqa:mc": 41.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 38.1 |
habanoz/tinyllama-oasst1-top1-instruct-full-lr1-5-v0.1 | main | e55b262cbd0ee52f7a4cbda136dbf1a027987c47 | {
"arc:challenge": 32.8,
"hellaswag": 58.2,
"hendrycksTest": 26,
"truthfulqa:mc": 38.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 38.8 |
Deci/DeciCoder-1b | main | af2ef45ef8cbe82eb7eb4074f260412bc14c7b11 | {
"arc:challenge": 21.2,
"hellaswag": 31.1,
"hendrycksTest": 24.3,
"truthfulqa:mc": 47.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.9 |
Andron00e/YetAnother_Open-Llama-3B-LoRA | main | 52c5cb0178831908ed0571f1750fcb0f0fb125f9 | {
"arc:challenge": 25.9,
"hellaswag": 25.8,
"hendrycksTest": 24.7,
"truthfulqa:mc": null
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | null |
Vmware/open-llama-7b-v2-open-instruct | main | b8fbe09571a71603ab517fe897a1281005060b62 | {
"arc:challenge": 39.8,
"hellaswag": 70.3,
"hendrycksTest": 35.2,
"truthfulqa:mc": 39.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 46.2 |
gpt2-medium | main | f65d4965d1221eff2bcf34f53a2ba12120e18f24 | {
"arc:challenge": 27,
"hellaswag": 40.2,
"hendrycksTest": 26.6,
"truthfulqa:mc": 40.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 33.7 |
TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T | main | f62ecb34ea0d4acea9d896040a4616a9538e2f36 | {
"arc:challenge": 30.3,
"hellaswag": 54.8,
"hendrycksTest": 26.5,
"truthfulqa:mc": 36.1
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 36.9 |
TinyLlama/TinyLlama-1.1B-Chat-v0.6 | main | bf9ae1c8bf026667e6f810768de259bb4a7f4777 | {
"arc:challenge": 31.7,
"hellaswag": 55.8,
"hendrycksTest": 26,
"truthfulqa:mc": 34.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 37 |
ai-forever/rugpt3large_based_on_gpt2 | main | 8201db0de8deb68f25e7309db04d163b71970494 | {
"arc:challenge": 22.6,
"hellaswag": 32.8,
"hendrycksTest": 24.9,
"truthfulqa:mc": 43.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.9 |
beomi/KoAlpaca-KoRWKV-6B | main | 427ee72c4350f26de1b287a0c07b842e7d168dbc | {
"arc:challenge": 23.5,
"hellaswag": 31.6,
"hendrycksTest": 24.9,
"truthfulqa:mc": 39.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30 |
beomi/KoRWKV-6B | main | 541600070459baf0f1be9560181d5ceb77794085 | {
"arc:challenge": 22.1,
"hellaswag": 32.2,
"hendrycksTest": 24.7,
"truthfulqa:mc": 39
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 29.5 |
gpt2 | main | 11c5a3d5811f50298f278a704980280950aedb10 | {
"arc:challenge": 22.1,
"hellaswag": null,
"hendrycksTest": null,
"truthfulqa:mc": null
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 22.1 |
pszemraj/pythia-31m-simplewiki-2048 | main | 95d47818055661250b55144c7d9beaf05dc126d8 | {
"arc:challenge": 22.2,
"hellaswag": 25.6,
"hendrycksTest": 23.1,
"truthfulqa:mc": 49.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.1 |
pszemraj/pythia-31m-goodwiki-deduped-2048-scratch | main | 01a3cd918dd7c233bc0c3c0c948a9a462a5359d1 | {
"arc:challenge": 23.1,
"hellaswag": 25.7,
"hendrycksTest": 23.1,
"truthfulqa:mc": 51.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 30.8 |
bavest/fin-llama-33b-merged | main | 17114520801da7b9599fe7a9fdf238915713a59b | {
"arc:challenge": 65,
"hellaswag": 86.2,
"hendrycksTest": 58.7,
"truthfulqa:mc": 49.7
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 64.9 |
totally-not-an-llm/EverythingLM-13b-V2-16k | main | 943f932ae1ae462389e6d2db5273158530749fff | {
"arc:challenge": 58.7,
"hellaswag": 80.9,
"hendrycksTest": 49.7,
"truthfulqa:mc": 47.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59.2 |
totally-not-an-llm/EverythingLM-13b-V3-peft | main | 7a2eed5038addcf4fa3b8dd358b45eb96134e749 | {
"arc:challenge": 58.4,
"hellaswag": 81,
"hendrycksTest": 54.7,
"truthfulqa:mc": 53
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 61.8 |
totally-not-an-llm/EverythingLM-13b-16k | main | 8456a856a8b115b05e76a7d0d945853b10ac71e2 | {
"arc:challenge": 56.6,
"hellaswag": 80.6,
"hendrycksTest": 50.2,
"truthfulqa:mc": 47.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.7 |
totally-not-an-llm/PuddleJumper-13b | main | f3a8a475ff0c6ae37ac8ae0690980be11cac731a | {
"arc:challenge": 58.7,
"hellaswag": 81.2,
"hendrycksTest": 58.3,
"truthfulqa:mc": 56.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 63.6 |
totally-not-an-llm/EverythingLM-13b-V3-16k | main | 1de9244bfadb947f80872727f76790cbc76e7142 | {
"arc:challenge": 58.2,
"hellaswag": 80.1,
"hendrycksTest": 50.5,
"truthfulqa:mc": 45.2
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 58.5 |
dvruette/gpt-neox-20b-full-precision | main | 20b347273d90da7c2c9eb4c32d4173dba862a0d2 | {
"arc:challenge": 48.8,
"hellaswag": 74.4,
"hendrycksTest": 26.2,
"truthfulqa:mc": 36.9
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 46.6 |
dvruette/oasst-pythia-12b-6000-steps | main | e2ccc0ef8d1cc5ffc8b0e2e885f03ef50597ea8a | {
"arc:challenge": 45.4,
"hellaswag": 69.7,
"hendrycksTest": 26,
"truthfulqa:mc": 39.8
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 45.2 |
dvruette/oasst-pythia-6.9b-4000-steps | main | 0e201b6f344ac6382dda40d389e1c9144a87d027 | {
"arc:challenge": 41.6,
"hellaswag": 64.2,
"hendrycksTest": 26.3,
"truthfulqa:mc": 40.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 43.1 |
dvruette/oasst-llama-13b-1000-steps | main | d2cd599cc40db3370009f45d6caa7e486cb6d31f | {
"arc:challenge": 58.1,
"hellaswag": 81.5,
"hendrycksTest": 48.6,
"truthfulqa:mc": 36
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 56 |
dvruette/oasst-llama-13b-2-epochs | main | 0e3796192f7edf43968541b9454ea35da4a2b1c5 | {
"arc:challenge": 57.9,
"hellaswag": 82.4,
"hendrycksTest": 48.6,
"truthfulqa:mc": 47.3
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 59 |
dvruette/oasst-gpt-neox-20b-3000-steps | main | f0462a8b7908f61202d86e6a9a2996d8339363b5 | {
"arc:challenge": 46.4,
"hellaswag": 72.1,
"hendrycksTest": 26.2,
"truthfulqa:mc": 35.5
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 45 |
dvruette/llama-13b-pretrained | main | c28cc0cf5a1a1bf4de96b23d06b02129dca85eb9 | {
"arc:challenge": 56.3,
"hellaswag": 79.3,
"hendrycksTest": 47,
"truthfulqa:mc": 48.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 57.8 |
dvruette/oasst-pythia-12b-reference | main | c5a9b7fad884e6c45ce5d2ca551aa1c03db6865f | {
"arc:challenge": 43,
"hellaswag": 67.9,
"hendrycksTest": 28.3,
"truthfulqa:mc": 36.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 44 |
h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b-preview-300bt-v2 | main | fdc6ff469295d0aaabec8948525b70d6688728ac | {
"arc:challenge": 36.4,
"hellaswag": 61.4,
"hendrycksTest": 25,
"truthfulqa:mc": 37.6
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 40.1 |
h2oai/h2ogpt-gm-oasst1-multilang-1024-20b | main | b3a6bf4250a037c09e451344e2a4e987011b79de | {
"arc:challenge": 47.4,
"hellaswag": 72.6,
"hendrycksTest": 26.4,
"truthfulqa:mc": 34.4
} | 9ba100d35ce48d3d4c132947464c93c861932caa | 2023-11-23T17:28:23 | 45.2 |