Commit
•
473d8d4
1
Parent(s):
611791e
Add tsv_format data files
Browse files- README.md +13 -5
- dataset_infos.json +11 -21
- tsv_format/test-00000-of-00001.parquet +3 -0
- tsv_format/train-00000-of-00001.parquet +3 -0
- tsv_format/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -90,16 +90,16 @@ dataset_info:
|
|
90 |
dtype: string
|
91 |
splits:
|
92 |
- name: train
|
93 |
-
num_bytes:
|
94 |
num_examples: 23097
|
95 |
- name: test
|
96 |
-
num_bytes:
|
97 |
num_examples: 2126
|
98 |
- name: validation
|
99 |
-
num_bytes:
|
100 |
num_examples: 1304
|
101 |
-
download_size:
|
102 |
-
dataset_size:
|
103 |
configs:
|
104 |
- config_name: snli_format
|
105 |
data_files:
|
@@ -109,6 +109,14 @@ configs:
|
|
109 |
path: snli_format/test-*
|
110 |
- split: validation
|
111 |
path: snli_format/validation-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
112 |
---
|
113 |
|
114 |
# Dataset Card for "scitail"
|
|
|
90 |
dtype: string
|
91 |
splits:
|
92 |
- name: train
|
93 |
+
num_bytes: 4606527
|
94 |
num_examples: 23097
|
95 |
- name: test
|
96 |
+
num_bytes: 410267
|
97 |
num_examples: 2126
|
98 |
- name: validation
|
99 |
+
num_bytes: 260422
|
100 |
num_examples: 1304
|
101 |
+
download_size: 1836546
|
102 |
+
dataset_size: 5277216
|
103 |
configs:
|
104 |
- config_name: snli_format
|
105 |
data_files:
|
|
|
109 |
path: snli_format/test-*
|
110 |
- split: validation
|
111 |
path: snli_format/validation-*
|
112 |
+
- config_name: tsv_format
|
113 |
+
data_files:
|
114 |
+
- split: train
|
115 |
+
path: tsv_format/train-*
|
116 |
+
- split: test
|
117 |
+
path: tsv_format/test-*
|
118 |
+
- split: validation
|
119 |
+
path: tsv_format/validation-*
|
120 |
---
|
121 |
|
122 |
# Dataset Card for "scitail"
|
dataset_infos.json
CHANGED
@@ -72,34 +72,30 @@
|
|
72 |
"size_in_bytes": 33203382
|
73 |
},
|
74 |
"tsv_format": {
|
75 |
-
"description": "The SciTail dataset is an entailment dataset created from multiple-choice science exams and web sentences. Each question
|
76 |
"citation": "inproceedings{scitail,\n Author = {Tushar Khot and Ashish Sabharwal and Peter Clark},\n Booktitle = {AAAI},\n Title = {{SciTail}: A Textual Entailment Dataset from Science Question Answering},\n Year = {2018}\n}\n",
|
77 |
"homepage": "https://allenai.org/data/scitail",
|
78 |
"license": "",
|
79 |
"features": {
|
80 |
"premise": {
|
81 |
"dtype": "string",
|
82 |
-
"id": null,
|
83 |
"_type": "Value"
|
84 |
},
|
85 |
"hypothesis": {
|
86 |
"dtype": "string",
|
87 |
-
"id": null,
|
88 |
"_type": "Value"
|
89 |
},
|
90 |
"label": {
|
91 |
"dtype": "string",
|
92 |
-
"id": null,
|
93 |
"_type": "Value"
|
94 |
}
|
95 |
},
|
96 |
-
"supervised_keys": null,
|
97 |
"builder_name": "scitail",
|
|
|
98 |
"config_name": "tsv_format",
|
99 |
"version": {
|
100 |
"version_str": "1.1.0",
|
101 |
"description": "",
|
102 |
-
"datasets_version_to_prepare": null,
|
103 |
"major": 1,
|
104 |
"minor": 1,
|
105 |
"patch": 0
|
@@ -107,32 +103,26 @@
|
|
107 |
"splits": {
|
108 |
"train": {
|
109 |
"name": "train",
|
110 |
-
"num_bytes":
|
111 |
"num_examples": 23097,
|
112 |
-
"dataset_name":
|
113 |
},
|
114 |
"test": {
|
115 |
"name": "test",
|
116 |
-
"num_bytes":
|
117 |
"num_examples": 2126,
|
118 |
-
"dataset_name":
|
119 |
},
|
120 |
"validation": {
|
121 |
"name": "validation",
|
122 |
-
"num_bytes":
|
123 |
"num_examples": 1304,
|
124 |
-
"dataset_name":
|
125 |
-
}
|
126 |
-
},
|
127 |
-
"download_checksums": {
|
128 |
-
"http://data.allenai.org.s3.amazonaws.com/downloads/SciTailV1.1.zip": {
|
129 |
-
"num_bytes": 14174621,
|
130 |
-
"checksum": "3fccd37350a94ca280b75998568df85fc2fc62843a3198d644fcbf858e6943d5"
|
131 |
}
|
132 |
},
|
133 |
-
"download_size":
|
134 |
-
"dataset_size":
|
135 |
-
"size_in_bytes":
|
136 |
},
|
137 |
"dgem_format": {
|
138 |
"description": "The SciTail dataset is an entailment dataset created from multiple-choice science exams and web sentences. Each question \nand the correct answer choice are converted into an assertive statement to form the hypothesis. We use information \nretrieval to obtain relevant text from a large text corpus of web sentences, and use these sentences as a premise P. We \ncrowdsource the annotation of such premise-hypothesis pair as supports (entails) or not (neutral), in order to create \nthe SciTail dataset. The dataset contains 27,026 examples with 10,101 examples with entails label and 16,925 examples \nwith neutral label\n",
|
|
|
72 |
"size_in_bytes": 33203382
|
73 |
},
|
74 |
"tsv_format": {
|
75 |
+
"description": "The SciTail dataset is an entailment dataset created from multiple-choice science exams and web sentences. Each question\nand the correct answer choice are converted into an assertive statement to form the hypothesis. We use information\nretrieval to obtain relevant text from a large text corpus of web sentences, and use these sentences as a premise P. We\ncrowdsource the annotation of such premise-hypothesis pair as supports (entails) or not (neutral), in order to create\nthe SciTail dataset. The dataset contains 27,026 examples with 10,101 examples with entails label and 16,925 examples\nwith neutral label\n",
|
76 |
"citation": "inproceedings{scitail,\n Author = {Tushar Khot and Ashish Sabharwal and Peter Clark},\n Booktitle = {AAAI},\n Title = {{SciTail}: A Textual Entailment Dataset from Science Question Answering},\n Year = {2018}\n}\n",
|
77 |
"homepage": "https://allenai.org/data/scitail",
|
78 |
"license": "",
|
79 |
"features": {
|
80 |
"premise": {
|
81 |
"dtype": "string",
|
|
|
82 |
"_type": "Value"
|
83 |
},
|
84 |
"hypothesis": {
|
85 |
"dtype": "string",
|
|
|
86 |
"_type": "Value"
|
87 |
},
|
88 |
"label": {
|
89 |
"dtype": "string",
|
|
|
90 |
"_type": "Value"
|
91 |
}
|
92 |
},
|
|
|
93 |
"builder_name": "scitail",
|
94 |
+
"dataset_name": "scitail",
|
95 |
"config_name": "tsv_format",
|
96 |
"version": {
|
97 |
"version_str": "1.1.0",
|
98 |
"description": "",
|
|
|
99 |
"major": 1,
|
100 |
"minor": 1,
|
101 |
"patch": 0
|
|
|
103 |
"splits": {
|
104 |
"train": {
|
105 |
"name": "train",
|
106 |
+
"num_bytes": 4606527,
|
107 |
"num_examples": 23097,
|
108 |
+
"dataset_name": null
|
109 |
},
|
110 |
"test": {
|
111 |
"name": "test",
|
112 |
+
"num_bytes": 410267,
|
113 |
"num_examples": 2126,
|
114 |
+
"dataset_name": null
|
115 |
},
|
116 |
"validation": {
|
117 |
"name": "validation",
|
118 |
+
"num_bytes": 260422,
|
119 |
"num_examples": 1304,
|
120 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
121 |
}
|
122 |
},
|
123 |
+
"download_size": 1836546,
|
124 |
+
"dataset_size": 5277216,
|
125 |
+
"size_in_bytes": 7113762
|
126 |
},
|
127 |
"dgem_format": {
|
128 |
"description": "The SciTail dataset is an entailment dataset created from multiple-choice science exams and web sentences. Each question \nand the correct answer choice are converted into an assertive statement to form the hypothesis. We use information \nretrieval to obtain relevant text from a large text corpus of web sentences, and use these sentences as a premise P. We \ncrowdsource the annotation of such premise-hypothesis pair as supports (entails) or not (neutral), in order to create \nthe SciTail dataset. The dataset contains 27,026 examples with 10,101 examples with entails label and 16,925 examples \nwith neutral label\n",
|
tsv_format/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2b4b8b5e258a30fe7d1f7861ad7154f1ebaf8f085f5e051db5e22352cf7ca96
|
3 |
+
size 162166
|
tsv_format/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:35ffcef823e42135a4fcee1b5ecb7c951e99f97b6f51c9363a23b537d41fb5d3
|
3 |
+
size 1574550
|
tsv_format/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7342d7d9c3f0c90b904b5fcfa37b909ed77fc3f9f0c4b87618d7718469f55b56
|
3 |
+
size 99830
|