Dataset Preview
Full Screen Viewer
Full Screen
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
The dataset generation failed because of a cast error
Error code: DatasetGenerationCastError Exception: DatasetGenerationCastError Message: An error occurred while generating the dataset All the data files must have the same columns, but at some point there are 1 new columns ({'job_id'}) This happened while the json dataset builder was generating data using hf://datasets/hallucinations-leaderboard/requests/EleutherAI/gpt-neo-1.3B_eval_request_False_False_False.json (at revision 7ee1a68af5f9531a3c2eb2ac39f5833e73a02e35) Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations) Traceback: Traceback (most recent call last): File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1869, in _prepare_split_single writer.write_table(table) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/arrow_writer.py", line 580, in write_table pa_table = table_cast(pa_table, self._schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2292, in table_cast return cast_table_to_schema(table, schema) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/table.py", line 2240, in cast_table_to_schema raise CastError( datasets.table.CastError: Couldn't cast model: string base_model: string revision: string private: bool status: string job_id: string weight_type: string precision: string model_type: string submitted_time: timestamp[s] license: string likes: int64 params: double to {'model': Value(dtype='string', id=None), 'base_model': Value(dtype='string', id=None), 'revision': Value(dtype='string', id=None), 'private': Value(dtype='bool', id=None), 'precision': Value(dtype='string', id=None), 'weight_type': Value(dtype='string', id=None), 'status': Value(dtype='string', id=None), 'submitted_time': Value(dtype='timestamp[s]', id=None), 'model_type': Value(dtype='string', id=None), 'likes': Value(dtype='int64', id=None), 'params': Value(dtype='float64', id=None), 'license': Value(dtype='string', id=None)} because column names don't match During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1392, in compute_config_parquet_and_info_response parquet_operations = convert_to_parquet(builder) File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1041, in convert_to_parquet builder.download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 924, in download_and_prepare self._download_and_prepare( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 999, in _download_and_prepare self._prepare_split(split_generator, **prepare_split_kwargs) File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1740, in _prepare_split for job_id, done, content in self._prepare_split_single( File "/src/services/worker/.venv/lib/python3.9/site-packages/datasets/builder.py", line 1871, in _prepare_split_single raise DatasetGenerationCastError.from_cast_error( datasets.exceptions.DatasetGenerationCastError: An error occurred while generating the dataset All the data files must have the same columns, but at some point there are 1 new columns ({'job_id'}) This happened while the json dataset builder was generating data using hf://datasets/hallucinations-leaderboard/requests/EleutherAI/gpt-neo-1.3B_eval_request_False_False_False.json (at revision 7ee1a68af5f9531a3c2eb2ac39f5833e73a02e35) Please either edit the data files to have matching columns, or separate them into different configurations (see docs at https://hf.co/docs/hub/datasets-manual-configuration#multiple-configurations)
Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
model
string | base_model
string | revision
string | private
bool | precision
string | weight_type
string | status
string | submitted_time
unknown | model_type
string | likes
int64 | params
float64 | license
string | job_id
string |
---|---|---|---|---|---|---|---|---|---|---|---|---|
01-ai/Yi-1.5-9B-Chat | main | false | bfloat16 | Original | FINISHED | "2024-05-27T17:06:08" | π¬ : chat models (RLHF, DPO, IFT, ...) | 80 | 8.829 | apache-2.0 | null |
|
BarraHome/zephyr-dpo-v2 | unsloth/zephyr-sft-bnb-4bit | main | false | float16 | Original | FINISHED | "2024-02-05T03:18:41" | πΆ : fine-tuned | 0 | 7.242 | mit | null |
DeepMount00/Llama-3-8b-Ita | main | false | bfloat16 | Original | FINISHED | "2024-05-17T15:15:42" | πΆ : fine-tuned on domain-specific datasets | 17 | 8.03 | llama3 | null |
|
EleutherAI/gpt-j-6b | main | false | float32 | Original | FINISHED | "2023-12-03T18:26:55" | pretrained | 1,316 | 6 | apache-2.0 | null |
|
EleutherAI/gpt-neo-1.3B | main | false | float32 | Original | FINISHED | "2023-09-09T10:52:17" | pretrained | 206 | 1.366 | mit | 506054 |
|
EleutherAI/gpt-neo-125m | main | false | float32 | Original | FINISHED | "2023-09-09T10:52:17" | pretrained | 132 | 0.15 | mit | 504945 |
|
EleutherAI/gpt-neo-2.7B | main | false | float32 | Original | FINISHED | "2023-09-09T10:52:17" | pretrained | 361 | 2.718 | mit | 460931 |
|
EleutherAI/llemma_7b | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:12" | pretrained | 54 | 7 | llama2 | null |
|
HuggingFaceH4/mistral-7b-sft-beta | main | false | float32 | Original | FINISHED | "2023-12-11T15:58:50" | fine-tuned | 15 | 7 | mit | null |
|
HuggingFaceH4/zephyr-7b-alpha | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:00" | fine-tuned | 986 | 7.242 | mit | null |
|
HuggingFaceH4/zephyr-7b-beta | main | false | float32 | Original | FINISHED | "2023-12-03T18:26:39" | fine-tuned | 973 | 7.242 | mit | null |
|
KnutJaegersberg/Qwen-14B-Llamafied | main | false | bfloat16 | Original | FINISHED | "2024-01-30T15:47:05" | π’ : pretrained | 1 | 14 | other | null |
|
KnutJaegersberg/internlm-20b-llama | main | false | bfloat16 | Original | RUNNING | "2024-01-30T15:47:15" | π’ : pretrained | 0 | 20 | other | null |
|
KoboldAI/GPT-J-6B-Janeway | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:31" | pretrained | 11 | 6 | mit | null |
|
KoboldAI/LLAMA2-13B-Holodeck-1 | main | false | float32 | Original | FINISHED | "2023-12-11T15:44:36" | pretrained | 20 | 13.016 | other | null |
|
KoboldAI/OPT-13B-Erebus | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:26" | pretrained | 168 | 13 | other | null |
|
KoboldAI/OPT-13B-Nerys-v2 | main | false | float32 | Original | FINISHED | "2023-12-07T22:48:11" | pretrained | 9 | 13 | other | null |
|
KoboldAI/OPT-2.7B-Erebus | main | false | float32 | Original | FINISHED | "2023-12-11T15:28:21" | pretrained | 32 | 2.7 | other | null |
|
KoboldAI/OPT-6.7B-Erebus | main | false | float32 | Original | FINISHED | "2023-12-07T22:48:16" | pretrained | 88 | 6.7 | other | null |
|
KoboldAI/OPT-6B-nerys-v2 | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:41" | pretrained | 21 | 6 | other | null |
|
KoboldAI/fairseq-dense-13B-Janeway | main | false | float32 | Original | FINISHED | "2023-12-11T15:19:11" | pretrained | 10 | 13 | mit | null |
|
LeoLM/leo-hessianai-7b-chat | main | false | float32 | Original | FINISHED | "2023-12-11T15:29:23" | instruction-tuned | 11 | 7 | null | null |
|
LeoLM/leo-hessianai-7b | main | false | float32 | Original | FINISHED | "2023-12-11T15:52:44" | pretrained | 32 | 7 | null | null |
|
NeverSleep/Noromaid-v0.4-Mixtral-Instruct-8x7b-Zloss | main | false | float16 | Original | RUNNING | "2024-01-29T17:05:05" | πΆ : fine-tuned | 10 | 46.703 | cc-by-nc-4.0 | null |
|
Nexusflow/NexusRaven-V2-13B | main | false | bfloat16 | Original | FINISHED | "2024-01-29T16:55:48" | πΆ : fine-tuned | 344 | 13 | other | null |
|
NotAiLOL/Yi-1.5-dolphin-9B | main | false | bfloat16 | Original | RUNNING | "2024-05-17T20:47:37" | πΆ : fine-tuned on domain-specific datasets | 0 | 8.829 | apache-2.0 | null |
|
NousResearch/Llama-2-13b-hf | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:45" | pretrained | 64 | 13.016 | null | null |
|
NousResearch/Llama-2-7b-chat-hf | main | false | float32 | Original | FINISHED | "2023-12-03T18:26:34" | instruction-tuned | 63 | 6.738 | null | null |
|
NousResearch/Llama-2-7b-hf | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:05" | pretrained | 92 | 6.738 | null | null |
|
NousResearch/Nous-Hermes-Llama2-13b | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:19" | pretrained | 251 | 13 | mit | null |
|
NousResearch/Nous-Hermes-llama-2-7b | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:18" | pretrained | 54 | 6.738 | mit | null |
|
NousResearch/Yarn-Mistral-7b-128k | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:34" | pretrained | 470 | 7 | apache-2.0 | null |
|
Open-Orca/Mistral-7B-OpenOrca | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:13" | pretrained | 485 | 7 | apache-2.0 | null |
|
Open-Orca/Mistral-7B-SlimOrca | main | false | float32 | Original | FINISHED | "2023-12-11T15:53:45" | pretrained | 21 | 7 | apache-2.0 | null |
|
Open-Orca/OpenOrca-Platypus2-13B | main | false | float32 | Original | FINISHED | "2023-12-11T15:20:13" | pretrained | 219 | 13 | cc-by-nc-4.0 | null |
|
Q-bert/MetaMath-Cybertron-Starling | main | false | bfloat16 | Original | RUNNING | "2024-01-30T15:57:50" | πΆ : fine-tuned | 38 | 7.242 | cc-by-nc-4.0 | null |
|
Steelskull/Umbra-v2.1-MoE-4x10.7 | main | false | bfloat16 | Original | RUNNING | "2024-01-29T13:48:43" | πΆ : fine-tuned | 1 | 36.099 | apache-2.0 | null |
|
TheBloke/Falcon-180B-Chat-GPTQ | main | false | float32 | Original | RUNNING | "2023-12-07T22:44:49" | fine-tuned | 66 | 197.968 | unknown | null |
|
TheBloke/Llama-2-13B-chat-AWQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:40" | instruction-tuned | 12 | 2.026 | llama2 | null |
|
TheBloke/Llama-2-13B-chat-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:31" | instruction-tuned | 309 | 16.232 | llama2 | null |
|
TheBloke/Llama-2-13B-fp16 | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:30" | pretrained | 53 | 13 | null | null |
|
TheBloke/Llama-2-70B-Chat-AWQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:44:51" | fine-tuned | 9 | 9.684 | llama2 | null |
|
TheBloke/Llama-2-70B-Chat-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:08" | fine-tuned | 232 | 72.816 | llama2 | null |
|
TheBloke/Llama-2-70B-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:16" | fine-tuned | 78 | 72.816 | llama2 | null |
|
TheBloke/Llama-2-7B-Chat-AWQ | main | false | float32 | Original | FINISHED | "2023-12-11T15:32:24" | fine-tuned | 8 | 1.129 | llama2 | null |
|
TheBloke/Llama-2-7B-Chat-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-03T18:26:23" | fine-tuned | 185 | 9.048 | llama2 | null |
|
TheBloke/Llama-2-7B-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:48:03" | fine-tuned | 65 | 9.048 | llama2 | null |
|
TheBloke/Mistral-7B-OpenOrca-AWQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:42" | fine-tuned | 31 | 1.196 | apache-2.0 | null |
|
TheBloke/Mistral-7B-OpenOrca-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:01" | fine-tuned | 83 | 9.592 | apache-2.0 | null |
|
TheBloke/Mythalion-13B-AWQ | main | false | float32 | Original | FINISHED | "2023-12-11T15:54:46" | fine-tuned | 4 | 2.026 | llama2 | null |
|
TheBloke/MythoMax-L2-13B-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:52" | fine-tuned | 97 | 16.232 | other | null |
|
TheBloke/Upstage-Llama-2-70B-instruct-v2-AWQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:48:32" | instruction-tuned | 0 | 9.684 | llama2 | null |
|
TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:12" | fine-tuned | 267 | 16.224 | other | null |
|
TheBloke/Wizard-Vicuna-30B-Uncensored-GPTQ | main | false | float32 | Original | RUNNING | "2023-12-07T22:47:06" | fine-tuned | 444 | 35.584 | other | null |
|
TheBloke/Wizard-Vicuna-7B-Uncensored-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:38" | fine-tuned | 120 | 9.04 | other | null |
|
TheBloke/orca_mini_v3_7B-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-11T15:35:27" | fine-tuned | 9 | 9.048 | other | null |
|
TheBloke/zephyr-7B-alpha-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:37" | fine-tuned | 25 | 9.592 | mit | null |
|
TheBloke/zephyr-7B-beta-AWQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:40" | fine-tuned | 26 | 1.196 | mit | null |
|
TheBloke/zephyr-7B-beta-GPTQ | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:21" | fine-tuned | 44 | 9.592 | mit | null |
|
TinyLlama/TinyLlama-1.1B-Chat-v0.3 | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:56" | pretrained | 33 | 1.1 | apache-2.0 | null |
|
TinyLlama/TinyLlama-1.1B-Chat-v0.6 | main | false | float32 | Original | FINISHED | "2023-12-11T15:21:15" | pretrained | 23 | 1.1 | apache-2.0 | null |
|
TinyLlama/TinyLlama-1.1B-intermediate-step-715k-1.5T | main | false | float32 | Original | FINISHED | "2023-12-11T15:42:34" | pretrained | 54 | 1.1 | apache-2.0 | null |
|
TinyLlama/TinyLlama-1.1B-intermediate-step-955k-token-2T | main | false | float32 | Original | FINISHED | "2023-12-11T14:35:43" | pretrained | 26 | 1.1 | apache-2.0 | null |
|
UjjwalP/testnic1 | main | false | float32 | Original | RUNNING | "2024-04-11T06:49:35" | πΆ : fine-tuned on domain-specific datasets | 0 | 0.582 | apache-2.0 | null |
|
aari1995/germeo-7b-laser | main | false | bfloat16 | Original | FINISHED | "2024-01-29T12:54:01" | π¦ : RL-tuned | 2 | 7.242 | apache-2.0 | null |
|
abacusai/Smaug-Llama-3-70B-Instruct | main | false | bfloat16 | Original | RUNNING | "2024-05-17T19:02:43" | π¬ : chat models (RLHF, DPO, IFT, ...) | 0 | 70.554 | llama2 | null |
|
ai-forever/mGPT | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:29" | pretrained | 190 | 0 | apache-2.0 | null |
|
ai4bharat/Airavata | main | false | float32 | Original | RUNNING | "2024-01-30T06:55:04" | πΆ : fine-tuned | 17 | 6.87 | llama2 | null |
|
augtoma/qCammel-70-x | main | false | float32 | Original | RUNNING | "2023-12-07T22:47:53" | pretrained | 22 | 0 | other | null |
|
berkeley-nest/Starling-LM-7B-alpha | main | false | float32 | Original | FINISHED | "2023-12-11T15:18:10" | pretrained | 335 | 7.242 | cc-by-nc-4.0 | null |
|
bigscience/bloom-1b1 | main | false | float32 | Original | FINISHED | "2023-12-07T22:47:14" | pretrained | 44 | 1.065 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloom-1b7 | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:16" | pretrained | 103 | 1.722 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloom-3b | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:26" | pretrained | 70 | 3.003 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloom-560m | main | false | float32 | Original | FINISHED | "2023-12-03T18:19:48" | pretrained | 285 | 0.559 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloom-7b1 | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:33" | pretrained | 148 | 7.069 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloom | main | false | float32 | Original | RUNNING | "2023-12-07T22:47:50" | pretrained | 4,218 | 176.247 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloomz-3b | main | false | float32 | Original | FINISHED | "2023-12-07T22:48:19" | pretrained | 72 | 3.003 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloomz-560m | main | false | float32 | Original | FINISHED | "2023-12-03T18:26:48" | pretrained | 85 | 0.559 | bigscience-bloom-rail-1.0 | null |
|
bigscience/bloomz-7b1 | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:46" | pretrained | 115 | 7 | bigscience-bloom-rail-1.0 | null |
|
cognitivecomputations/WizardLM-33B-V1.0-Uncensored | main | false | float16 | Original | RUNNING | "2024-01-30T15:58:52" | πΆ : fine-tuned | 56 | 33 | other | null |
|
cognitivecomputations/dolphin-2.9.1-yi-1.5-34b | main | false | bfloat16 | Original | FINISHED | "2024-05-25T05:41:27" | πΆ : fine-tuned on domain-specific datasets | 28 | 34.389 | apache-2.0 | null |
|
ddeshler/HerLlama-7B-slerp | main | false | bfloat16 | Original | FINISHED | "2024-08-27T19:58:30" | π€ : base merges and moerges | 0 | 7.242 | apache-2.0 | null |
|
deepseek-ai/deepseek-llm-67b-chat | main | false | float16 | Original | RUNNING | "2024-01-30T15:59:40" | πΆ : fine-tuned | 142 | 67 | other | null |
|
distilbert/distilroberta-base | main | false | float32 | Original | FINISHED | "2024-05-10T01:33:43" | π’ : pretrained | 119 | 0.083 | apache-2.0 | null |
|
ehartford/dolphin-2.1-mistral-7b | main | false | float32 | Original | FINISHED | "2023-12-11T15:39:30" | pretrained | 233 | 7 | apache-2.0 | null |
|
ehartford/dolphin-2.2.1-mistral-7b | main | false | float32 | Original | FINISHED | "2023-12-07T22:45:44" | fine-tuned | 118 | 7 | apache-2.0 | null |
|
facebook/bart-base | main | false | float32 | Original | RUNNING | "2024-03-15T17:02:42" | 137 | 0.139 | apache-2.0 | null |
||
facebook/opt-13b | main | false | float32 | Original | FINISHED | "2023-12-07T22:46:09" | pretrained | 60 | 13 | other | null |
|
facebook/xglm-564M | main | false | float32 | Original | FINISHED | "2023-12-11T15:33:26" | pretrained | 32 | 0.564 | mit | null |
|
fearlessdots/WizardLM-2-7B-abliterated | main | false | bfloat16 | Original | FINISHED | "2024-05-25T13:37:35" | πΆ : fine-tuned on domain-specific datasets | 1 | 7.242 | apache-2.0 | null |
|
garage-bAInd/Platypus2-7B | main | false | float32 | Original | FINISHED | "2023-12-11T15:55:47" | pretrained | 4 | 7 | cc-by-nc-sa-4.0 | null |
|
google/flan-t5-base | main | false | float32 | Original | RUNNING | "2024-03-12T18:59:10" | pretrained | 598 | 0.248 | apache-2.0 | null |
|
google/flan-t5-large | main | false | float32 | Original | RUNNING | "2024-03-12T18:59:17" | pretrained | 428 | 0.783 | apache-2.0 | null |
|
google/flan-t5-xxl | main | false | float32 | Original | RUNNING | "2024-03-12T18:59:01" | pretrained | 1,074 | 11.267 | apache-2.0 | null |
|
google/gemma-1.1-7b-it | main | false | bfloat16 | Original | RUNNING | "2024-05-16T16:23:41" | π¬ : chat models (RLHF, DPO, IFT, ...) | 230 | 8.538 | gemma | null |
|
google/gemma-2b | main | false | float32 | Original | FINISHED | "2024-03-12T19:31:34" | pretrained | 548 | 2.506 | other | null |
|
google/gemma-7b-it | main | false | float32 | Original | RUNNING | "2024-03-12T19:31:30" | pretrained | 918 | 8.538 | other | null |
|
google/gemma-7b | main | false | bfloat16 | Original | FINISHED | "2024-05-16T16:23:49" | π’ : pretrained | 2,880 | 8.538 | gemma | null |
|
google/gemma-7b | main | false | float32 | Original | FINISHED | "2024-03-12T19:31:27" | pretrained | 2,214 | 8.538 | other | null |
|
h2oai/h2ogpt-4096-llama2-13b-chat | main | false | float32 | Original | FINISHED | "2023-12-03T18:27:23" | instruction-tuned | 10 | 13.016 | llama2 | null |
End of preview.
README.md exists but content is empty.
Use the Edit dataset card button to edit it.
- Downloads last month
- 4