davzoku commited on
Commit
f0afadd
1 Parent(s): a2a4f46

Convert dataset to Parquet

Browse files
README.md CHANGED
@@ -24,6 +24,7 @@ task_ids:
24
  paperswithcode_id: quac
25
  pretty_name: Question Answering in Context
26
  dataset_info:
 
27
  features:
28
  - name: dialogue_id
29
  dtype: string
@@ -65,16 +66,22 @@ dataset_info:
65
  sequence: string
66
  - name: answer_starts
67
  sequence: int32
68
- config_name: plain_text
69
  splits:
70
  - name: train
71
- num_bytes: 58174754
72
  num_examples: 11567
73
  - name: validation
74
- num_bytes: 7375938
75
  num_examples: 1000
76
- download_size: 77043986
77
- dataset_size: 65550692
 
 
 
 
 
 
 
78
  ---
79
 
80
  # Dataset Card for Question Answering in Context
 
24
  paperswithcode_id: quac
25
  pretty_name: Question Answering in Context
26
  dataset_info:
27
+ config_name: plain_text
28
  features:
29
  - name: dialogue_id
30
  dtype: string
 
66
  sequence: string
67
  - name: answer_starts
68
  sequence: int32
 
69
  splits:
70
  - name: train
71
+ num_bytes: 58174602
72
  num_examples: 11567
73
  - name: validation
74
+ num_bytes: 7375862
75
  num_examples: 1000
76
+ download_size: 34925990
77
+ dataset_size: 65550464
78
+ configs:
79
+ - config_name: plain_text
80
+ data_files:
81
+ - split: train
82
+ path: plain_text/train-*
83
+ - split: validation
84
+ path: plain_text/validation-*
85
  ---
86
 
87
  # Dataset Card for Question Answering in Context
dataset_infos.json CHANGED
@@ -1 +1,127 @@
1
- {"plain_text": {"description": "Question Answering in Context is a dataset for modeling, understanding,\nand participating in information seeking dialog. Data instances consist\nof an interactive dialog between two crowd workers: (1) a student who\nposes a sequence of freeform questions to learn as much as possible\nabout a hidden Wikipedia text, and (2) a teacher who answers the questions\nby providing short excerpts (spans) from the text. QuAC introduces\nchallenges not found in existing machine comprehension datasets: its\nquestions are often more open-ended, unanswerable, or only meaningful\nwithin the dialog context.\n", "citation": "@inproceedings{choi-etal-2018-quac,\ntitle = \"QUAC: Question answering in context\",\nabstract = \"We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total). The dialogs involve two crowd workers: (1) a student who poses a sequence of freeform questions to learn as much as possible about a hidden Wikipedia text, and (2) a teacher who answers the questions by providing short excerpts from the text. QuAC introduces challenges not found in existing machine comprehension datasets: its questions are often more open-ended, unanswerable, or only meaningful within the dialog context, as we show in a detailed qualitative evaluation. We also report results for a number of reference models, including a recently state-of-the-art reading comprehension architecture extended to model dialog context. Our best model underperforms humans by 20 F1, suggesting that there is significant room for future work on this data. Dataset, baseline, and leaderboard available at http://quac.ai.\",\nauthor = \"Eunsol Choi and He He and Mohit Iyyer and Mark Yatskar and Yih, {Wen Tau} and Yejin Choi and Percy Liang and Luke Zettlemoyer\",\nyear = \"2018\",\nlanguage = \"English (US)\",\nseries = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018\",\npublisher = \"Association for Computational Linguistics\",\npages = \"2174--2184\",\neditor = \"Ellen Riloff and David Chiang and Julia Hockenmaier and Jun'ichi Tsujii\",\nbooktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018\",\nnote = \"2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018 ; Conference date: 31-10-2018 Through 04-11-2018\",\n}\n", "homepage": "https://quac.ai/", "license": "MIT", "features": {"dialogue_id": {"dtype": "string", "id": null, "_type": "Value"}, "wikipedia_page_title": {"dtype": "string", "id": null, "_type": "Value"}, "background": {"dtype": "string", "id": null, "_type": "Value"}, "section_title": {"dtype": "string", "id": null, "_type": "Value"}, "context": {"dtype": "string", "id": null, "_type": "Value"}, "turn_ids": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "questions": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "followups": {"feature": {"num_classes": 3, "names": ["y", "n", "m"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}, "yesnos": {"feature": {"num_classes": 3, "names": ["y", "n", "x"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}, "answers": {"feature": {"texts": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer_starts": {"feature": {"dtype": "int32", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}, "orig_answers": {"texts": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "answer_starts": {"feature": {"dtype": "int32", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}}, "post_processed": null, "supervised_keys": null, "builder_name": "quac", "config_name": "plain_text", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 58174754, "num_examples": 11567, "dataset_name": "quac"}, "validation": {"name": "validation", "num_bytes": 7375938, "num_examples": 1000, "dataset_name": "quac"}}, "download_checksums": {"https://s3.amazonaws.com/my89public/quac/train_v0.2.json": {"num_bytes": 68114819, "checksum": "ff5cca5a2e4b4d1cb5b5ced68b9fce88394ef6d93117426d6d4baafbcc05c56a"}, "https://s3.amazonaws.com/my89public/quac/val_v0.2.json": {"num_bytes": 8929167, "checksum": "09e622916280ba04c9352acb1bc5bbe80f11a2598f6f34e934c51d9e6570f378"}}, "download_size": 77043986, "post_processing_size": null, "dataset_size": 65550692, "size_in_bytes": 142594678}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "plain_text": {
3
+ "description": "Question Answering in Context is a dataset for modeling, understanding,\nand participating in information seeking dialog. Data instances consist\nof an interactive dialog between two crowd workers: (1) a student who\nposes a sequence of freeform questions to learn as much as possible\nabout a hidden Wikipedia text, and (2) a teacher who answers the questions\nby providing short excerpts (spans) from the text. QuAC introduces\nchallenges not found in existing machine comprehension datasets: its\nquestions are often more open-ended, unanswerable, or only meaningful\nwithin the dialog context.\n",
4
+ "citation": "@inproceedings{choi-etal-2018-quac,\ntitle = \"QUAC: Question answering in context\",\nabstract = \"We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions in total). The dialogs involve two crowd workers: (1) a student who poses a sequence of freeform questions to learn as much as possible about a hidden Wikipedia text, and (2) a teacher who answers the questions by providing short excerpts from the text. QuAC introduces challenges not found in existing machine comprehension datasets: its questions are often more open-ended, unanswerable, or only meaningful within the dialog context, as we show in a detailed qualitative evaluation. We also report results for a number of reference models, including a recently state-of-the-art reading comprehension architecture extended to model dialog context. Our best model underperforms humans by 20 F1, suggesting that there is significant room for future work on this data. Dataset, baseline, and leaderboard available at http://quac.ai.\",\nauthor = \"Eunsol Choi and He He and Mohit Iyyer and Mark Yatskar and Yih, {Wen Tau} and Yejin Choi and Percy Liang and Luke Zettlemoyer\",\nyear = \"2018\",\nlanguage = \"English (US)\",\nseries = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018\",\npublisher = \"Association for Computational Linguistics\",\npages = \"2174--2184\",\neditor = \"Ellen Riloff and David Chiang and Julia Hockenmaier and Jun'ichi Tsujii\",\nbooktitle = \"Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018\",\nnote = \"2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018 ; Conference date: 31-10-2018 Through 04-11-2018\",\n}\n",
5
+ "homepage": "https://quac.ai/",
6
+ "license": "MIT",
7
+ "features": {
8
+ "dialogue_id": {
9
+ "dtype": "string",
10
+ "_type": "Value"
11
+ },
12
+ "wikipedia_page_title": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "background": {
17
+ "dtype": "string",
18
+ "_type": "Value"
19
+ },
20
+ "section_title": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ },
24
+ "context": {
25
+ "dtype": "string",
26
+ "_type": "Value"
27
+ },
28
+ "turn_ids": {
29
+ "feature": {
30
+ "dtype": "string",
31
+ "_type": "Value"
32
+ },
33
+ "_type": "Sequence"
34
+ },
35
+ "questions": {
36
+ "feature": {
37
+ "dtype": "string",
38
+ "_type": "Value"
39
+ },
40
+ "_type": "Sequence"
41
+ },
42
+ "followups": {
43
+ "feature": {
44
+ "names": [
45
+ "y",
46
+ "n",
47
+ "m"
48
+ ],
49
+ "_type": "ClassLabel"
50
+ },
51
+ "_type": "Sequence"
52
+ },
53
+ "yesnos": {
54
+ "feature": {
55
+ "names": [
56
+ "y",
57
+ "n",
58
+ "x"
59
+ ],
60
+ "_type": "ClassLabel"
61
+ },
62
+ "_type": "Sequence"
63
+ },
64
+ "answers": {
65
+ "feature": {
66
+ "texts": {
67
+ "feature": {
68
+ "dtype": "string",
69
+ "_type": "Value"
70
+ },
71
+ "_type": "Sequence"
72
+ },
73
+ "answer_starts": {
74
+ "feature": {
75
+ "dtype": "int32",
76
+ "_type": "Value"
77
+ },
78
+ "_type": "Sequence"
79
+ }
80
+ },
81
+ "_type": "Sequence"
82
+ },
83
+ "orig_answers": {
84
+ "texts": {
85
+ "feature": {
86
+ "dtype": "string",
87
+ "_type": "Value"
88
+ },
89
+ "_type": "Sequence"
90
+ },
91
+ "answer_starts": {
92
+ "feature": {
93
+ "dtype": "int32",
94
+ "_type": "Value"
95
+ },
96
+ "_type": "Sequence"
97
+ }
98
+ }
99
+ },
100
+ "builder_name": "parquet",
101
+ "dataset_name": "quac",
102
+ "config_name": "plain_text",
103
+ "version": {
104
+ "version_str": "1.1.0",
105
+ "major": 1,
106
+ "minor": 1,
107
+ "patch": 0
108
+ },
109
+ "splits": {
110
+ "train": {
111
+ "name": "train",
112
+ "num_bytes": 58174602,
113
+ "num_examples": 11567,
114
+ "dataset_name": null
115
+ },
116
+ "validation": {
117
+ "name": "validation",
118
+ "num_bytes": 7375862,
119
+ "num_examples": 1000,
120
+ "dataset_name": null
121
+ }
122
+ },
123
+ "download_size": 34925990,
124
+ "dataset_size": 65550464,
125
+ "size_in_bytes": 100476454
126
+ }
127
+ }
plain_text/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:067c94373d1a622339a73f16032d8060bb5dd989cd8ce2b3647363a8bc1d5b5f
3
+ size 31324530
plain_text/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9701bfc0cbb60cd9b90e2419effc59f5b8cd5057b987f861ec7b0c25c90c6b7a
3
+ size 3601460