model
stringlengths 4
89
| revision
stringclasses 1
value | model_sha
stringlengths 0
40
| results
dict | commit
stringlengths 40
40
| date
timestamp[ns] | score
float64 21.8
83
⌀ |
---|---|---|---|---|---|---|
TheBloke/Kimiko-v2-13B-fp16 | main | 0fed305667508e50330e71a2d43e9cee5ea73783 | {
"arc:challenge": 61,
"hellaswag": 83.3,
"hendrycksTest": 55.2,
"truthfulqa:mc": 40.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60 |
TheBloke/wizard-vicuna-13B-GPTQ | main | 936a51c0219744d7a9598d0c65a7d18e01660601 | {
"arc:challenge": 28.7,
"hellaswag": 25.9,
"hendrycksTest": 25.8,
"truthfulqa:mc": 48.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.2 |
TheBloke/CodeLlama-34B-Python-fp16 | main | 875f9d97fb6c9619d8867887dd1d80918ff0f593 | {
"arc:challenge": 38.1,
"hellaswag": 34.8,
"hendrycksTest": 32.9,
"truthfulqa:mc": 43.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 37.4 |
TheBloke/manticore-13b-chat-pyg-GPTQ | main | 923f27245d13058c9c1b3ab0eab6c6c93ffc162e | {
"arc:challenge": 57.8,
"hellaswag": 81.1,
"hendrycksTest": 47.6,
"truthfulqa:mc": 47.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.6 |
TheBloke/Airoboros-L2-70B-2.1-GPTQ | main | 23ed580cb77ebaee49ea11eb4538fd3ab3795b76 | {
"arc:challenge": 70.4,
"hellaswag": 86.5,
"hendrycksTest": 68.9,
"truthfulqa:mc": 55.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70.3 |
TheBloke/Airoboros-L2-13B-2.1-GPTQ | main | d90d96e40b9359cb5c35e6b6c8f0eb24896e827b | {
"arc:challenge": 59,
"hellaswag": 81.7,
"hendrycksTest": 53.2,
"truthfulqa:mc": 44.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.6 |
TheBloke/gpt4-x-vicuna-13B-HF | main | a247577c882940e0c6b040fe8239d760c0d10d40 | {
"arc:challenge": 53.4,
"hellaswag": 80.1,
"hendrycksTest": 51.2,
"truthfulqa:mc": 53.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.6 |
TheBloke/Wizard-Vicuna-30B-Superhot-8K-fp16 | main | 062fe5409861d7386279fb534b435be39c88ceaf | {
"arc:challenge": 26.2,
"hellaswag": 33,
"hendrycksTest": 23.5,
"truthfulqa:mc": 47.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.6 |
TheBloke/Wizard-Vicuna-7B-Uncensored-HF | main | b802f1b4401d0b2242137160c20cc11b9ffd3a4c | {
"arc:challenge": 53.4,
"hellaswag": 78.8,
"hendrycksTest": 37.1,
"truthfulqa:mc": 43.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.2 |
TheBloke/Wizard-Vicuna-13B-Uncensored-HF | main | fff9ac7f0e2e7b340f2301f5f089d989fc03be67 | {
"arc:challenge": 59,
"hellaswag": 81.9,
"hendrycksTest": 47.9,
"truthfulqa:mc": 51.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.1 |
TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ | main | 56a82ece7a9309189561a590e8f4d2fe0d4be92b | {
"arc:challenge": 61.1,
"hellaswag": 82.4,
"hendrycksTest": 56.5,
"truthfulqa:mc": 49.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 62.5 |
TheBloke/orca_mini_v3_7B-GPTQ | main | 06ddd48cd904907e3c73d2dfe47d28626053598b | {
"arc:challenge": 30.1,
"hellaswag": 26,
"hendrycksTest": 24.3,
"truthfulqa:mc": 48.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.2 |
TheBloke/OpenOrca-Platypus2-13B-GPTQ | main | 0fa9a56066656fbc94e3ec088bc900fd1d4d38e8 | {
"arc:challenge": 62.5,
"hellaswag": 82.7,
"hendrycksTest": 58.6,
"truthfulqa:mc": 51.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 63.9 |
TheBloke/guanaco-33B-GPTQ | main | 8e42e031bfc8be3bbf31dc546d7c51fb991ff6e0 | {
"arc:challenge": 28.2,
"hellaswag": 26.3,
"hendrycksTest": 24.9,
"truthfulqa:mc": 49
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.1 |
TheBloke/robin-33B-v2-GPTQ | main | 4c2588d65302e9ca634548ed81e8650fb2975686 | {
"arc:challenge": 27.7,
"hellaswag": 26.3,
"hendrycksTest": 23.5,
"truthfulqa:mc": 49.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.8 |
TheBloke/WizardLM-33B-V1.0-Uncensored-GPTQ | main | 1c65902c620fcdf6b9c8e36ce17f21360e186a1e | {
"arc:challenge": 27.4,
"hellaswag": 26,
"hendrycksTest": 25.8,
"truthfulqa:mc": 48.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32 |
TheBloke/wizard-vicuna-13B-HF | main | 12dc8aacb474522ae2a83c18cb0fdf0907987f8f | {
"arc:challenge": 54.7,
"hellaswag": 79.2,
"hendrycksTest": 48.9,
"truthfulqa:mc": 49.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 58.1 |
TheBloke/vicuna-13B-1.1-HF | main | 8c71dbe9221e83d2ec72e4dc08beccfc78b563c0 | {
"arc:challenge": 52.7,
"hellaswag": 80.1,
"hendrycksTest": 51.9,
"truthfulqa:mc": 52.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.2 |
TheBloke/WizardLM-13B-V1-1-SuperHOT-8K-GPTQ | main | 085eb5cd394f30d72bf5efcf83a580e87264b3e8 | {
"arc:challenge": 57,
"hellaswag": 80.3,
"hendrycksTest": 47.1,
"truthfulqa:mc": 53.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.5 |
TheBloke/Chinese-Alpaca-33B-SuperHOT-8K-fp16 | main | a55ce761bace8be6d17c357c57ef927751afd40c | {
"arc:challenge": 26.8,
"hellaswag": 29.6,
"hendrycksTest": 24.1,
"truthfulqa:mc": 47.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32 |
TheBloke/gpt4-alpaca-lora-30b-HF | main | 3c8007467a081dc72ae09b9d358416b056b38920 | {
"arc:challenge": 64.8,
"hellaswag": 85.7,
"hendrycksTest": 58.5,
"truthfulqa:mc": 52.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 65.3 |
TheBloke/tulu-30B-fp16 | main | 37c3655676c37662f60c68dacfce3f0e861be846 | {
"arc:challenge": 60,
"hellaswag": 83.4,
"hendrycksTest": 56.1,
"truthfulqa:mc": 45.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 61.2 |
TheBloke/airoboros-33B-gpt4-1-4-SuperHOT-8K-fp16 | main | 53fdac1cdb8a37647e5dbe4199bc3fb70e617fce | {
"arc:challenge": 26,
"hellaswag": 30.7,
"hendrycksTest": 23.6,
"truthfulqa:mc": 47.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.1 |
TheBloke/EverythingLM-13B-16K-GPTQ | main | f14d3df05577f3e1ac35e2c4ec32ce0d39b97508 | {
"arc:challenge": 29.3,
"hellaswag": 26.2,
"hendrycksTest": 25.4,
"truthfulqa:mc": 48.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.4 |
TheBloke/Llama-2-7B-GPTQ | main | ecd7ab9f6adc36ecbe0d751eeea0d90ae1863c3b | {
"arc:challenge": 52,
"hellaswag": 77.6,
"hendrycksTest": 44,
"truthfulqa:mc": 39.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.2 |
TheBloke/Kimiko-13B-fp16 | main | 27868769e2d6b1af46337f0997c71b0577952a3d | {
"arc:challenge": 59.2,
"hellaswag": 82.4,
"hendrycksTest": 55.8,
"truthfulqa:mc": 39.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.3 |
TheBloke/CodeLlama-34B-Instruct-fp16 | main | a4d0ce949de4d5b5f74691641efb5b70736a32a8 | {
"arc:challenge": 40.8,
"hellaswag": 35.7,
"hendrycksTest": 39.7,
"truthfulqa:mc": 44.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 40.1 |
TheBloke/WizardLM-70B-V1.0-GPTQ | main | c234d7c9c0fd26efb55757fdbfb604d549539fe0 | {
"arc:challenge": 63.8,
"hellaswag": 83.8,
"hendrycksTest": 63.7,
"truthfulqa:mc": 54.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 66.4 |
TheBloke/CodeLlama-13B-Python-fp16 | main | 442282f4207442b828953a72c51a919c332cba5c | {
"arc:challenge": 33.2,
"hellaswag": 44.5,
"hendrycksTest": 25.9,
"truthfulqa:mc": 44
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 36.9 |
TheBloke/Manticore-13B-Chat-Pyg-Guanaco-SuperHOT-8K-GPTQ | main | bd3c66e626c81de4977f197e1534bd3dfa2f569d | {
"arc:challenge": 52.8,
"hellaswag": 79.6,
"hendrycksTest": 39.8,
"truthfulqa:mc": 52.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.2 |
TheBloke/medalpaca-13B-GPTQ-4bit | main | 12190f743a19e91dfe1f5c77abc0c1bf486073dd | {
"arc:challenge": 29.4,
"hellaswag": 26.3,
"hendrycksTest": 25.4,
"truthfulqa:mc": 49.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.6 |
TheBloke/OpenOrcaxOpenChat-Preview2-13B-GPTQ | main | ec9eb4f471b5bb6a7e5e505369628586c0c72252 | {
"arc:challenge": 61.3,
"hellaswag": 82.1,
"hendrycksTest": 57.8,
"truthfulqa:mc": 50.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 62.8 |
TheBloke/llama-2-70b-Guanaco-QLoRA-fp16 | main | 54b0e39d5e9aee7b323f50b0a26db15295c3d5c9 | {
"arc:challenge": 68.3,
"hellaswag": 88.3,
"hendrycksTest": 70.2,
"truthfulqa:mc": 55.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70.6 |
TheBloke/tulu-13B-fp16 | main | 532aeb363b0ceee155b3cf9479ef635b797cee7c | {
"arc:challenge": 53.9,
"hellaswag": 80.7,
"hendrycksTest": 53.2,
"truthfulqa:mc": 43.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57.9 |
TheBloke/Lemur-70B-Chat-v1-GPTQ | main | 12499165df1785f50df3e95940406032776401ea | {
"arc:challenge": 65.3,
"hellaswag": 84.4,
"hendrycksTest": 64.7,
"truthfulqa:mc": 57.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 67.9 |
TheBloke/wizardLM-7B-HF | main | a8e22531a48cece989e670f539eb18ebd2dbd0cf | {
"arc:challenge": 50.3,
"hellaswag": 75.3,
"hendrycksTest": 38.1,
"truthfulqa:mc": 45.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 52.3 |
TheBloke/Llama-2-13B-GPTQ | main | b7db471d1789802a3a8e3b93cdd66a9f046f17c3 | {
"arc:challenge": 57.3,
"hellaswag": 81.6,
"hendrycksTest": 54.8,
"truthfulqa:mc": 36.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57.6 |
TheBloke/dromedary-65b-lora-HF | main | 3fa4546259d6bbd6b5d637484c325ab19181a73c | {
"arc:challenge": 61.6,
"hellaswag": 82.5,
"hendrycksTest": 63.1,
"truthfulqa:mc": 38.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 61.5 |
TheBloke/landmark-attention-llama7b-fp16 | main | bf8bdcb0c30cceb0ceda33cf5fde683807e39a58 | {
"arc:challenge": 47.4,
"hellaswag": 65.8,
"hendrycksTest": 31.6,
"truthfulqa:mc": 42.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 46.8 |
TheBloke/stable-vicuna-13B-HF | main | 2b099b2be0dafb2606ae9808c0f6183fe4bff7bc | {
"arc:challenge": 53.3,
"hellaswag": 78.5,
"hendrycksTest": 50.3,
"truthfulqa:mc": 48.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57.6 |
TheBloke/WizardLM-13B-V1.1-GPTQ | main | 9df807ac64034bc6e7387326689d6e39656ce5e0 | {
"arc:challenge": 58.5,
"hellaswag": 80.7,
"hendrycksTest": 49.6,
"truthfulqa:mc": 54.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.8 |
TheBloke/robin-13B-v2-fp16 | main | f4dd8fc4440ed84fcf3ff1122f2b7f6024cca29d | {
"arc:challenge": 56.5,
"hellaswag": 80.4,
"hendrycksTest": 48.8,
"truthfulqa:mc": 50.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.1 |
TheBloke/VicUnlocked-alpaca-65B-QLoRA-fp16 | main | 6cdacfda96970aa144e316b108ab9bc17c99a573 | {
"arc:challenge": 65.6,
"hellaswag": 85.2,
"hendrycksTest": 63.1,
"truthfulqa:mc": 52.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 66.6 |
TheBloke/koala-7B-HF | main | d102fe3b68f1a5a50d547e4fd1c8b33b783c993b | {
"arc:challenge": 47.1,
"hellaswag": 73.6,
"hendrycksTest": 25.5,
"truthfulqa:mc": 46
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 48 |
TheBloke/Nous-Hermes-13B-SuperHOT-8K-fp16 | main | b407c1ece029ad5693d38e6e0931e9482962ed15 | {
"arc:challenge": 55.3,
"hellaswag": 81.9,
"hendrycksTest": 48.2,
"truthfulqa:mc": 51.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.1 |
TheBloke/tulu-7B-fp16 | main | 8a026683f79119643f4007da4e9155c7849792cc | {
"arc:challenge": 50.2,
"hellaswag": 77,
"hendrycksTest": 47.6,
"truthfulqa:mc": 41.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54.1 |
TheBloke/Vicuna-33B-1-3-SuperHOT-8K-fp16 | main | 0b6484697d5cca5baa534b882dcad8101add8cda | {
"arc:challenge": 25.4,
"hellaswag": 34.6,
"hendrycksTest": 23.6,
"truthfulqa:mc": 46.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.6 |
TheBloke/wizardLM-13B-1.0-fp16 | main | b79733805e98e668ff9a459975c259881b1b8014 | {
"arc:challenge": 57.3,
"hellaswag": 80.9,
"hendrycksTest": 52.9,
"truthfulqa:mc": 50.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.4 |
TheBloke/Genz-70b-GPTQ | main | 7d38987a43d2445b193db99a029a264b39dc6c8e | {
"arc:challenge": 71.1,
"hellaswag": 87.6,
"hendrycksTest": 70.3,
"truthfulqa:mc": 62.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 72.8 |
TheBloke/Project-Baize-v2-7B-GPTQ | main | 5dc039834e1ea42ac334458b2e3090fe3705cc59 | {
"arc:challenge": 46,
"hellaswag": 73.4,
"hendrycksTest": 35.5,
"truthfulqa:mc": 39.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 48.7 |
TheBloke/airoboros-13B-HF | main | 9219b61a0e8bc880e4cd0f8bebc48a97ee0950c7 | {
"arc:challenge": 58.3,
"hellaswag": 81,
"hendrycksTest": 50,
"truthfulqa:mc": 51.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60.2 |
TheBloke/Platypus2-70B-Instruct-GPTQ | main | 4a44568aadd8a4babfa5549cf33e6e84cbae7ab8 | {
"arc:challenge": 71.2,
"hellaswag": 87.6,
"hendrycksTest": 69.9,
"truthfulqa:mc": 62.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 72.8 |
TheBloke/guanaco-7B-HF | main | 293c24105fa15afa127a2ec3905fdc2a0a3a6dac | {
"arc:challenge": 53,
"hellaswag": 80.1,
"hendrycksTest": 35.3,
"truthfulqa:mc": 39.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 51.9 |
TheBloke/koala-13B-HF | main | b20f96a0171ce4c0fa27d6048215ebe710521587 | {
"arc:challenge": 53,
"hellaswag": 77.6,
"hendrycksTest": 45.3,
"truthfulqa:mc": 50.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56.5 |
chaoyi-wu/MedLLaMA_13B | main | 893557ef32f98cd01deb1c5d063be6d640ffa657 | {
"arc:challenge": 54.3,
"hellaswag": 78.5,
"hendrycksTest": 46.4,
"truthfulqa:mc": 40.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54.9 |
Voicelab/trurl-2-13b | main | c8b2bbc7a570a9ea67928674695a4e7dff017d66 | {
"arc:challenge": 60.1,
"hellaswag": 80.2,
"hendrycksTest": 78.6,
"truthfulqa:mc": 45.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 66.2 |
Voicelab/trurl-2-7b | main | e26ca5f157c60fc527170cc04db7fc0ea04ad26f | {
"arc:challenge": 53.4,
"hellaswag": 75.3,
"hendrycksTest": 50,
"truthfulqa:mc": 45.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 56 |
THUDM/chatglm2-6b | main | 162b620e3078b03eefff94eb5f762d4093425fb5 | {
"arc:challenge": 38.8,
"hellaswag": 59,
"hendrycksTest": 46.7,
"truthfulqa:mc": 48.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 48.2 |
Brillibits/Instruct_Llama70B_Dolly15k | main | 45444ac60488594e0700e6c7313ff444b4468240 | {
"arc:challenge": 68.3,
"hellaswag": 87.2,
"hendrycksTest": 69.5,
"truthfulqa:mc": 46.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 67.9 |
joehuangx/spatial-vicuna-7b-v1.5-LoRA | main | dc71924cfb214b91461d35178e6ea6fef7946f13 | {
"arc:challenge": 50.8,
"hellaswag": 74.6,
"hendrycksTest": 48.1,
"truthfulqa:mc": 49.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.7 |
openthaigpt/openthaigpt-1.0.0-alpha-7b-chat-ckpt-hf | main | cdffb3488c5cb1a9aa5039a6b3bc72af24827db0 | {
"arc:challenge": 50.9,
"hellaswag": 74.9,
"hendrycksTest": 40,
"truthfulqa:mc": 47.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.2 |
breadlicker45/dough-instruct-base-001 | main | 3e1b0bf0a887feeb342982eee4f6d8041772a7dd | {
"arc:challenge": 23.9,
"hellaswag": 24.8,
"hendrycksTest": 23.1,
"truthfulqa:mc": 53.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.3 |
Tincando/fiction_story_generator | main | 377b080cf96e10d50289aa3e1fd79c330265f45a | {
"arc:challenge": 23.3,
"hellaswag": 28.7,
"hendrycksTest": 26.7,
"truthfulqa:mc": 43.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30.6 |
GOAT-AI/GOAT-7B-Community | main | a7073a0f5142ce04aaa1603b0812b358f62a8de8 | {
"arc:challenge": 48.8,
"hellaswag": 74.6,
"hendrycksTest": 49.6,
"truthfulqa:mc": 42.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.9 |
l3utterfly/open-llama-3b-v2-layla | main | 465669ddafad25393ac3cfe94d3726cced112b30 | {
"arc:challenge": 38.2,
"hellaswag": 66.4,
"hendrycksTest": 28.6,
"truthfulqa:mc": 44.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 44.4 |
SaylorTwift/gpt2_test | main | ef61310a16ffda93bf8f6132e02658482ffc2bcc | {
"arc:challenge": 21.8,
"hellaswag": 31.6,
"hendrycksTest": 25.9,
"truthfulqa:mc": 40.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30 |
Lazycuber/L2-7b-Guanaco-Uncensored | main | 9d49378c69c00113cf7f6e66d1ddb9d9b003dddc | {
"arc:challenge": 50.6,
"hellaswag": 77,
"hendrycksTest": 48.9,
"truthfulqa:mc": 43.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55 |
Lazycuber/pyg-instruct-wizardlm | main | f00ef7a7b0cc6f02af2a11ac764270dfd61b9e2f | {
"arc:challenge": 41,
"hellaswag": 66.7,
"hendrycksTest": 26.3,
"truthfulqa:mc": 31.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 41.5 |
Lazycuber/Janemalion-6B | main | e72ae3ec110121115b1ae6c2e5fb3995997a2d96 | {
"arc:challenge": 42.4,
"hellaswag": 68.4,
"hendrycksTest": 28.3,
"truthfulqa:mc": 34.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 43.4 |
MrNJK/gpt2-xl-sft | main | 53250831436460254b7ee9afc4014d4d3156b372 | {
"arc:challenge": 30,
"hellaswag": 49.2,
"hendrycksTest": 25.6,
"truthfulqa:mc": 38.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 35.9 |
malhajar/Platypus2-70B-instruct-4bit-gptq | main | 2aa2f5646e496b3cd9b510681ba2c5081bde821f | {
"arc:challenge": 29,
"hellaswag": 26,
"hendrycksTest": 23.5,
"truthfulqa:mc": 49.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32 |
pillowtalks-ai/delta13b | main | 83fa0860990df1db35550f973ba4306449e35412 | {
"arc:challenge": 52.7,
"hellaswag": 80.1,
"hendrycksTest": 51.9,
"truthfulqa:mc": 52.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.2 |
ethzanalytics/pythia-31m | main | 8a3c2f1555de8a3c53d67d73b5d0d53a66a6c6c2 | {
"arc:challenge": 20,
"hellaswag": 26.3,
"hendrycksTest": 24.3,
"truthfulqa:mc": 50.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30.2 |
Rachneet/gpt2-xl-alpaca | main | a1a19acc0ef161bfa35f460c15ed3015595714d8 | {
"arc:challenge": 26.8,
"hellaswag": 43.8,
"hendrycksTest": 26.3,
"truthfulqa:mc": 39.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 34.1 |
Panchovix/airoboros-33b-gpt4-1.2-SuperHOT-8k | main | 47c14f699cbbc9bd24458edd86eb70d87552b623 | {
"arc:challenge": 24.7,
"hellaswag": 31.2,
"hendrycksTest": 23.1,
"truthfulqa:mc": 47.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 31.6 |
Panchovix/WizardLM-33B-V1.0-Uncensored-SuperHOT-8k | main | b6d0002b10d43ab48aa14e365d9e7b40655ec160 | {
"arc:challenge": 25.4,
"hellaswag": 32,
"hendrycksTest": 23.4,
"truthfulqa:mc": 47
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32 |
HyperbeeAI/Tulpar-7b-v1 | main | 719d8e1eb4a820f01e0a92ef6220d041964bb472 | {
"arc:challenge": 57,
"hellaswag": 79.7,
"hendrycksTest": 51.3,
"truthfulqa:mc": 51.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60 |
quantumaikr/QuantumLM | main | 9058130b416355b37f5f78777748aa56d98a4da0 | {
"arc:challenge": 55.8,
"hellaswag": 79.7,
"hendrycksTest": 54.2,
"truthfulqa:mc": 46.7
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 59.1 |
quantumaikr/KoreanLM-hf | main | a7261e7ae6ee76c78e1ba1ac8c59bcc3e0868bf9 | {
"arc:challenge": 51.5,
"hellaswag": 76.8,
"hendrycksTest": 40.6,
"truthfulqa:mc": 44.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 53.3 |
quantumaikr/quantumairk-llama-2-70B-instruct | main | 94ff2fcafd507b08e953f70806ec671ec3d17b15 | {
"arc:challenge": 70.3,
"hellaswag": 87.1,
"hendrycksTest": 70.5,
"truthfulqa:mc": 54.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70.6 |
quantumaikr/llama-2-7b-hf-guanaco-1k | main | bdb57c5c992872ced47f48cb2177a5fa159f926a | {
"arc:challenge": 51.6,
"hellaswag": 76.7,
"hendrycksTest": 47.4,
"truthfulqa:mc": 44.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 55.1 |
quantumaikr/QuantumLM-70B-hf | main | e13dd23ae5e611e959b6c8d5bc47bf4fd37cd9d7 | {
"arc:challenge": 59.5,
"hellaswag": 83,
"hendrycksTest": 62.3,
"truthfulqa:mc": 53.4
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 64.6 |
quantumaikr/llama-2-70b-fb16-guanaco-1k | main | c317af1b593a4f91b0e79c7142ca75f1e8d65278 | {
"arc:challenge": 70.5,
"hellaswag": 87.3,
"hendrycksTest": 70.2,
"truthfulqa:mc": 57.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 71.4 |
quantumaikr/llama-2-70b-fb16-korean | main | fd57855006c15c4121feccab1cbeee8107de5b5a | {
"arc:challenge": 67.2,
"hellaswag": 86.8,
"hendrycksTest": 69.3,
"truthfulqa:mc": 56.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70 |
quantumaikr/open_llama_7b_hf | main | 41441cea58f963cfc4827da12ae5759e943151cb | {
"arc:challenge": 26.5,
"hellaswag": 26.9,
"hendrycksTest": 26.5,
"truthfulqa:mc": 49.5
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 32.4 |
quantumaikr/llama-2-70B-chat | main | d242fdbf800e388e6ee456578064cab5e057f987 | {
"arc:challenge": 67.6,
"hellaswag": 86.9,
"hendrycksTest": 69.2,
"truthfulqa:mc": 57.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70.2 |
quantumaikr/llama-2-70b-fb16-orca-chat-10k | main | 697aaeb8eb9905c9b25bebb736d1905444c774a6 | {
"arc:challenge": 68.1,
"hellaswag": 87.1,
"hendrycksTest": 69.2,
"truthfulqa:mc": 61.6
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 71.5 |
quantumaikr/QuantumLM-llama2-70B-Korean-LoRA | main | ea21456e999f6ce35da1cd88b8f62bb5770b985a | {
"arc:challenge": 70.6,
"hellaswag": 86.4,
"hendrycksTest": 69.4,
"truthfulqa:mc": 56.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 70.6 |
quantumaikr/QuantumLM-7B | main | f44998432fb90d88094ddf42e57ec458877a197f | {
"arc:challenge": 50.3,
"hellaswag": 76.1,
"hendrycksTest": 45.3,
"truthfulqa:mc": 46.3
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 54.5 |
huggingtweets/jerma985 | main | 816206ad02a397161be78dcb70eeda67e0c53132 | {
"arc:challenge": 21.7,
"hellaswag": 30.9,
"hendrycksTest": 26.6,
"truthfulqa:mc": 44
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 30.8 |
huggingtweets/gladosystem | main | 02a1bbcee7b584ace743b2fe4885cc0eaf2179ac | {
"arc:challenge": 24.4,
"hellaswag": 29.7,
"hendrycksTest": 23.2,
"truthfulqa:mc": 41.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 29.8 |
oh-yeontaek/llama-2-7B-LoRA-assemble | main | 72e866a96a2e9afc6527c8d757c69088c3a069c8 | {
"arc:challenge": 57.3,
"hellaswag": 78.8,
"hendrycksTest": 50.8,
"truthfulqa:mc": 53.2
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 60 |
oh-yeontaek/llama-2-70B-LoRA-assemble-v2 | main | 7feeb5b665ab1ecdfd9cc4fe45fadb86b7b91b5b | {
"arc:challenge": 71.8,
"hellaswag": 86.9,
"hendrycksTest": 69.4,
"truthfulqa:mc": 64.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 73.2 |
oh-yeontaek/llama-2-70B-LoRA-assemble | main | 91caffe08852dcbbdedd64786bd3b4ac0dcb2e96 | {
"arc:challenge": 71.8,
"hellaswag": 86.8,
"hendrycksTest": 69.4,
"truthfulqa:mc": 64.8
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 73.2 |
oh-yeontaek/llama-2-13B-LoRA-assemble | main | 85bb49d333dba4a08b051418663d16853ce30cee | {
"arc:challenge": 63.6,
"hellaswag": 83.5,
"hendrycksTest": 59.8,
"truthfulqa:mc": 56
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 65.7 |
teknium/OpenHermes-7B | main | 74edb1ad58d3d517ef46c4e2a31081084ecbc473 | {
"arc:challenge": 56.1,
"hellaswag": 78.3,
"hendrycksTest": 48.6,
"truthfulqa:mc": 45
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 57 |
teknium/OpenHermes-13B | main | f09d0fe655ad57cce9179b7b40ea6f81e07db18c | {
"arc:challenge": 60.2,
"hellaswag": 82.2,
"hendrycksTest": 56.2,
"truthfulqa:mc": 46
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 61.2 |
psmathur/model_007 | main | 0f5d81b13718a866cb078bd8762ab80a41972663 | {
"arc:challenge": 71.1,
"hellaswag": 87.7,
"hendrycksTest": 69,
"truthfulqa:mc": 63.1
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 72.7 |
psmathur/orca_mini_7b | main | 6ed0dca683685cb5b9e7df599f87d311f00ba6db | {
"arc:challenge": 43.9,
"hellaswag": 65.2,
"hendrycksTest": 30,
"truthfulqa:mc": 42
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 45.3 |
psmathur/orca_mini_v2_7b | main | 165850882991d7fa4eabab577a03ed84e0713bfa | {
"arc:challenge": 50.8,
"hellaswag": 76,
"hendrycksTest": 39.5,
"truthfulqa:mc": 43.9
} | 10ee91ffb0d887b2c4d3f1c5a32f9c4752237da3 | 2023-09-20T10:22:33 | 52.6 |