MicPie commited on
Commit
e2fdef0
1 Parent(s): 1af7d10

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +251 -0
README.md ADDED
@@ -0,0 +1,251 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ annotations_creators:
3
+ - no-annotation
4
+ language_creators:
5
+ - found
6
+ languages:
7
+ - en
8
+ license:
9
+ - apache-2.0
10
+ multilinguality:
11
+ - monolingual
12
+ pretty_name: AdapTable-wkdu.org
13
+ size_categories:
14
+ - 100K<n<1M
15
+ source_datasets: []
16
+ task_categories:
17
+ - multiple-choice
18
+ - question-answering
19
+ - zero-shot-classification
20
+ - text2text-generation
21
+ - table-question-answering
22
+ - text-generation
23
+ - text-classification
24
+ - tabular-classification
25
+ task_ids:
26
+ - multiple-choice-qa
27
+ - extractive-qa
28
+ - open-domain-qa
29
+ - closed-domain-qa
30
+ - closed-book-qa
31
+ - open-book-qa
32
+ - language-modeling
33
+ - multi-class-classification
34
+ - natural-language-inference
35
+ - topic-classification
36
+ - multi-label-classification
37
+ - tabular-multi-class-classification
38
+ - tabular-multi-label-classification
39
+ ---
40
+
41
+
42
+ # Dataset Card for "AdapTable-wkdu.org" - Dataset of Few-shot Tasks from Tables
43
+
44
+ ## Table of Contents
45
+ - [Dataset Description](#dataset-description)
46
+ - [Dataset Summary](#dataset-summary)
47
+ - [Supported Tasks](#supported-tasks-and-leaderboards)
48
+ - [Languages](#languages)
49
+ - [Dataset Structure](#dataset-structure)
50
+ - [Data Instances](#data-instances)
51
+ - [Data Fields](#data-instances)
52
+ - [Data Splits](#data-instances)
53
+ - [Dataset Creation](#dataset-creation)
54
+ - [Curation Rationale](#curation-rationale)
55
+ - [Source Data](#source-data)
56
+ - [Annotations](#annotations)
57
+ - [Personal and Sensitive Information](#personal-and-sensitive-information)
58
+ - [Considerations for Using the Data](#considerations-for-using-the-data)
59
+ - [Social Impact of Dataset](#social-impact-of-dataset)
60
+ - [Discussion of Biases](#discussion-of-biases)
61
+ - [Other Known Limitations](#other-known-limitations)
62
+ - [Additional Information](#additional-information)
63
+ - [Dataset Curators](#dataset-curators)
64
+ - [Licensing Information](#licensing-information)
65
+ - [Citation Information](#citation-information)
66
+
67
+ ## Dataset Description
68
+
69
+ - **Homepage:** [Needs More Information]
70
+ - **Repository:** https://github.com/JunShern/few-shot-pretraining
71
+ - **Paper:** Exploring Few-Shot Adaptation of Language Models with Tables
72
+ - **Leaderboard:** [Needs More Information]
73
+ - **Point of Contact:** [email protected], [email protected]
74
+
75
+ ### Dataset Summary
76
+
77
+ The AdapTable dataset consists of tables that naturally occur on the web and that are formatted as few-shot tasks for fine-tuning language models to improve their few-shot performance.
78
+
79
+ There are several dataset versions available:
80
+
81
+ * [AdapTable-full](https://huggingface.co/datasets/MicPie/adaptable_full): Starting from the initial WTC corpus of 50M tables, we apply our tables-to-tasks procedure to produce our resulting dataset, [AdapTable-full](https://huggingface.co/datasets/MicPie/adaptable_full), which comprises 413,350 tasks from 23,744 unique websites.
82
+
83
+ * [AdapTable-unique](https://huggingface.co/datasets/MicPie/adaptable_unique): This is the same as [AdapTable-full](https://huggingface.co/datasets/MicPie/adaptable_full) but filtered to have a maximum of one task per website. [AdapTable-unique](https://huggingface.co/datasets/MicPie/adaptable_unique) contains exactly 23,744 tasks from 23,744 websites.
84
+
85
+ * [AdapTable-5k](https://huggingface.co/datasets/MicPie/adaptable_5k): This dataset uses 5k random tables from the full dataset.
86
+
87
+ * AdapTable data subsets based on a manual human quality rating:
88
+ * [AdapTable-rated-low](https://huggingface.co/datasets/MicPie/adaptable_rated-low)
89
+ * [AdapTable-rated-medium](https://huggingface.co/datasets/MicPie/adaptable_rated-medium)
90
+ * [AdapTable-rated-high](https://huggingface.co/datasets/MicPie/adaptable_rated-high)
91
+
92
+ * AdapTable data subsets based on the website of origin:
93
+ * [AdapTable-baseball.fantasysports.yahoo.com](https://huggingface.co/datasets/MicPie/adaptable_baseball.fantasysports.yahoo.com)
94
+ * [AdapTable-bulbapedia.bulbagarden.net](https://huggingface.co/datasets/MicPie/adaptable_bulbapedia.bulbagarden.net)
95
+ * [AdapTable-cappex.com](https://huggingface.co/datasets/MicPie/adaptable_cappex.com)
96
+ * [AdapTable-cram.com](https://huggingface.co/datasets/MicPie/adaptable_cram.com)
97
+ * [AdapTable-dividend.com](https://huggingface.co/datasets/MicPie/adaptable_dividend.com)
98
+ * [AdapTable-dummies.com](https://huggingface.co/datasets/MicPie/adaptable_dummies.com)
99
+ * [AdapTable-en.wikipedia.org](https://huggingface.co/datasets/MicPie/adaptable_en.wikipedia.org)
100
+ * [AdapTable-ensembl.org](https://huggingface.co/datasets/MicPie/adaptable_ensembl.org)
101
+ * [AdapTable-gamefaqs.com](https://huggingface.co/datasets/MicPie/adaptable_gamefaqs.com)
102
+ * [AdapTable-mgoblog.com](https://huggingface.co/datasets/MicPie/adaptable_mgoblog.com)
103
+ * [AdapTable-mmo-champion.com](https://huggingface.co/datasets/MicPie/adaptable_mmo-champion.com)
104
+ * [AdapTable-msdn.microsoft.com](https://huggingface.co/datasets/MicPie/adaptable_msdn.microsoft.com)
105
+ * [AdapTable-phonearena.com](https://huggingface.co/datasets/MicPie/adaptable_phonearena.com)
106
+ * [AdapTable-sittercity.com](https://huggingface.co/datasets/MicPie/adaptable_sittercity.com)
107
+ * [AdapTable-sporcle.com](https://huggingface.co/datasets/MicPie/adaptable_sporcle.com)
108
+ * [AdapTable-studystack.com](https://huggingface.co/datasets/MicPie/adaptable_studystack.com)
109
+ * [AdapTable-support.google.com](https://huggingface.co/datasets/MicPie/adaptable_support.google.com)
110
+ * [AdapTable-w3.org](https://huggingface.co/datasets/MicPie/adaptable_w3.org)
111
+ * [AdapTable-wiki.openmoko.org](https://huggingface.co/datasets/MicPie/adaptable_wiki.openmoko.org)
112
+ * [AdapTable-wkdu.org](https://huggingface.co/datasets/MicPie/adaptable_wkdu.org)
113
+
114
+
115
+ * AdapTable data subsets based on clustering (for the clustering details please see our publication):
116
+ * [AdapTable-cluster00](https://huggingface.co/datasets/MicPie/adaptable_cluster00)
117
+ * [AdapTable-cluster01](https://huggingface.co/datasets/MicPie/adaptable_cluster01)
118
+ * [AdapTable-cluster02](https://huggingface.co/datasets/MicPie/adaptable_cluster02)
119
+ * [AdapTable-cluster03](https://huggingface.co/datasets/MicPie/adaptable_cluster03)
120
+ * [AdapTable-cluster04](https://huggingface.co/datasets/MicPie/adaptable_cluster04)
121
+ * [AdapTable-cluster05](https://huggingface.co/datasets/MicPie/adaptable_cluster05)
122
+ * [AdapTable-cluster06](https://huggingface.co/datasets/MicPie/adaptable_cluster06)
123
+ * [AdapTable-cluster07](https://huggingface.co/datasets/MicPie/adaptable_cluster07)
124
+ * [AdapTable-cluster08](https://huggingface.co/datasets/MicPie/adaptable_cluster08)
125
+ * [AdapTable-cluster09](https://huggingface.co/datasets/MicPie/adaptable_cluster09)
126
+ * [AdapTable-cluster10](https://huggingface.co/datasets/MicPie/adaptable_cluster10)
127
+ * [AdapTable-cluster11](https://huggingface.co/datasets/MicPie/adaptable_cluster11)
128
+ * [AdapTable-cluster12](https://huggingface.co/datasets/MicPie/adaptable_cluster12)
129
+ * [AdapTable-cluster13](https://huggingface.co/datasets/MicPie/adaptable_cluster13)
130
+ * [AdapTable-cluster14](https://huggingface.co/datasets/MicPie/adaptable_cluster14)
131
+ * [AdapTable-cluster15](https://huggingface.co/datasets/MicPie/adaptable_cluster15)
132
+ * [AdapTable-cluster16](https://huggingface.co/datasets/MicPie/adaptable_cluster16)
133
+ * [AdapTable-cluster17](https://huggingface.co/datasets/MicPie/adaptable_cluster17)
134
+ * [AdapTable-cluster18](https://huggingface.co/datasets/MicPie/adaptable_cluster18)
135
+ * [AdapTable-cluster19](https://huggingface.co/datasets/MicPie/adaptable_cluster19)
136
+ * [AdapTable-cluster20](https://huggingface.co/datasets/MicPie/adaptable_cluster20)
137
+ * [AdapTable-cluster21](https://huggingface.co/datasets/MicPie/adaptable_cluster21)
138
+ * [AdapTable-cluster22](https://huggingface.co/datasets/MicPie/adaptable_cluster22)
139
+ * [AdapTable-cluster23](https://huggingface.co/datasets/MicPie/adaptable_cluster23)
140
+ * [AdapTable-cluster24](https://huggingface.co/datasets/MicPie/adaptable_cluster24)
141
+ * [AdapTable-cluster25](https://huggingface.co/datasets/MicPie/adaptable_cluster25)
142
+ * [AdapTable-cluster26](https://huggingface.co/datasets/MicPie/adaptable_cluster26)
143
+ * [AdapTable-cluster27](https://huggingface.co/datasets/MicPie/adaptable_cluster27)
144
+ * [AdapTable-cluster28](https://huggingface.co/datasets/MicPie/adaptable_cluster28)
145
+ * [AdapTable-cluster29](https://huggingface.co/datasets/MicPie/adaptable_cluster29)
146
+ * [AdapTable-cluster-noise](https://huggingface.co/datasets/MicPie/adaptable_cluster-noise)
147
+
148
+
149
+
150
+ ### Supported Tasks and Leaderboards
151
+
152
+ Since the tables come from the web, the distribution of tasks and topics is very broad. The shape of our dataset is very wide, i.e., we have 1000's of tasks, while each task has only a few examples, compared to most current NLP datasets which are very deep, i.e., 10s of tasks with many examples. This implies that our dataset covers a broad range of potential tasks, e.g., multiple-choice, question-answering, table-question-answering, text-classification, etc.
153
+
154
+ The intended use of this dataset is to improve few-shot performance by fine-tuning/pre-training on our dataset.
155
+
156
+ ### Languages
157
+
158
+ English
159
+
160
+ ## Dataset Structure
161
+
162
+ ### Data Instances
163
+
164
+ Each table, i.e., task is represented as a json-lines file and consists of several few-shot examples. Each example is a dictionary containing a field 'task', which identifies the task, followed by an 'input', 'options', and 'output' field. The 'input' field contains several column elements of the same row in the table, while the 'output' field is a target which represents an individual column of the same row. Each task contains several such examples which can be concatenated as a few-shot task. In the case of multiple choice classification, the 'options' field contains the possible classes that a model needs to choose from.
165
+
166
+ There are also additional meta-data fields such as 'pageTitle', 'title', 'outputColName', 'url', 'wdcFile'.
167
+
168
+ ### Data Fields
169
+
170
+ 'task': task identifier
171
+
172
+ 'input': column elements of a specific row in the table.
173
+
174
+ 'options': for multiple choice classification, it provides the options to choose from.
175
+
176
+ 'output': target column element of the same row as input.
177
+
178
+ 'pageTitle': the title of the page containing the table.
179
+
180
+ 'outputColName': output column name
181
+
182
+ 'url': url to the website containing the table
183
+
184
+ 'wdcFile': WDC Web Table Corpus file
185
+
186
+ ### Data Splits
187
+
188
+ The AdapTable datasets do not come with additional data splits.
189
+
190
+ ## Dataset Creation
191
+
192
+ ### Curation Rationale
193
+
194
+ How do we convert tables to few-shot tasks?
195
+ Unlike unstructured text, structured data in the form of tables lends itself easily to the few-shot task format. Given a table where each row is an instance of a similar class and the columns describe the attributes of each instance, we can turn each row into a task example to predict one attribute given the others. When the table has more than one row, we instantly have multiple examples of this task by using each row as a single example, and thus each table becomes a few-shot dataset for a particular task.
196
+
197
+ The few-shot setting in this setup is significant: Tables often do not come with clear instructions for each field, so tasks may be underspecified if prompted in a zero-shot manner, but the intended task becomes clearer when examples are provided. This makes a good two-way match: The few-shot format is a perfect setup for table learning, and tables provide a natural dataset for few-shot training.
198
+
199
+ ### Source Data
200
+
201
+ #### Initial Data Collection and Normalization
202
+
203
+ The data processing pipeline is explained in detail in our publication.
204
+
205
+ #### Who are the source language producers?
206
+
207
+ The dataset is extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/).
208
+
209
+ ### Annotations
210
+
211
+ #### Annotation process
212
+
213
+ No manual annotation process used.
214
+ Only for the [AdapTable-rated-low](https://huggingface.co/datasets/MicPie/adaptable_rated-low), [AdapTable-rated-medium](https://huggingface.co/datasets/MicPie/adaptable_rated-medium), and [AdapTable-rated-high](https://huggingface.co/datasets/MicPie/adaptable_rated-high) manual annotations were carried out.
215
+
216
+ #### Who are the annotators?
217
+
218
+ People involved in the publication.
219
+
220
+ ### Personal and Sensitive Information
221
+
222
+ The data was extracted from [WDC Web Table Corpora](http://webdatacommons.org/webtables/), which in turn extracted tables from the [Common Crawl](https://commoncrawl.org/). We did not filter the data in any way. Thus any user identities or otherwise sensitive information (e.g., data that reveals racial or ethnic origins, sexual orientations, religious beliefs, political opinions or union memberships, or locations; financial or health data; biometric or genetic data; forms of government identification, such as social security numbers; criminal history, etc.) might be contained in our dataset.
223
+
224
+ ## Considerations for Using the Data
225
+
226
+ ### Social Impact of Dataset
227
+
228
+ The purpose of this dataset is to help develop models that are better at few-shot learning and have higher few-shot performance by fine-tuning few-shot tasks extracted from tables.
229
+
230
+ While tables have a similar structure to few-shot tasks and we do see an improved performance on few-shot tasks in our paper, we want to make clear that fine-tuning on tables also has its risks. First of all, since the tables are extracted from the web, they may contain user identities or otherwise sensitive information which a model might reveal at inference, or which could influence the learning process of a model in a negative way. Second, since tables are very diverse in nature, the model also trains on low-quality data or data with an unusual structure. While it is interesting that training on such data improves few-shot performance on downstream tasks, this could also imply that the model learns concepts that are very dissimilar to human concepts that would be useful for a certain downstream task. In other words, it is possible that the model learns weird things that are helpful on the evaluated downstream tasks, but might lead to bad out-of-distribution behavior.
231
+
232
+ ### Discussion of Biases
233
+
234
+ Since our dataset contains tables that are scraped from the web, it will also contain many toxic, racist, sexist, and otherwise harmful biases and texts. We have not run any analysis on the biases prevalent in our datasets. Neither have we explicitly filtered the content.
235
+ This implies that a model trained on our dataset will potentially reinforce harmful biases and toxic text that exist in our dataset.
236
+
237
+ ### Other Known Limitations
238
+
239
+ No additional known limitations.
240
+
241
+ ## Additional Information
242
+
243
+ ### Dataset Curators
244
+ Jun Shern Chan, Michael Pieler, Jonathan Jao, Jérémy Scheurer, Ethan Perez
245
+
246
+ ### Licensing Information
247
+ Apache 2.0
248
+
249
+ ### Citation Information
250
+
251
+ [Needs More Information]