rebel-dataset / dataset_infos.json
PereLluis13's picture
add dummy and infos
beb1d6c
raw
history blame
2.06 kB
{"REBEL": {"description": "REBEL is a silver dataset created for the paper REBEL: Relation Extraction By End-to-end Language generation\n", "citation": " @inproceedings{huguet-cabot-navigli-2021-rebel,\n title = \"REBEL: Relation Extraction By End-to-end Language generation\",\n author = \"Huguet Cabot, Pere-Llu{'\\i}s and\n Navigli, Roberto\",\n booktitle = \"Findings of the Association for Computational Linguistics: EMNLP 2021\",\n month = nov,\n year = \"2021\",\n address = \"Online and in the Barcel\u00f3 B\u00e1varo Convention Centre, Punta Cana, Dominican Republic\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://github.com/Babelscape/rebel/blob/main/docs/EMNLP_2021_REBEL__Camera_Ready_.pdf\",\n }\n", "homepage": "https://github.com/Babelscape/rebel", "license": "Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License (CC BY-NC-SA 4.0)", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "title": {"dtype": "string", "id": null, "_type": "Value"}, "context": {"dtype": "string", "id": null, "_type": "Value"}, "triplets": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "rebel", "config_name": "REBEL", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1079651403, "num_examples": 3120296, "dataset_name": "rebel"}, "validation": {"name": "validation", "num_bytes": 59712299, "num_examples": 172860, "dataset_name": "rebel"}, "test": {"name": "test", "num_bytes": 59909573, "num_examples": 173601, "dataset_name": "rebel"}}, "download_checksums": {"https://huggingface.co/datasets/Babelscape/rebel-dataset/resolve/main/rebel_dataset.zip": {"num_bytes": 1490017445, "checksum": "3e5a84a32b604b80617acc868ae3437e00f07e88244d72fbda44ccfbb3989980"}}, "download_size": 1490017445, "post_processing_size": null, "dataset_size": 1199273275, "size_in_bytes": 2689290720}}