Datasets:

Tasks:
Other
Languages:
English
ArXiv:
License:
File size: 14,480 Bytes
16f3142
 
 
 
 
ca07daa
16f3142
ca07daa
 
16f3142
 
 
 
 
 
 
 
 
 
 
 
 
 
7dc6e17
16f3142
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7dc6e17
16f3142
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7dc6e17
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
---
annotations_creators:
- crowdsourced
language_creators:
- crowdsourced
language:
- en
license:
- other
multilinguality:
- monolingual
paperswithcode_id: charades
pretty_name: Charades
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- other
task_ids:
- other
---

# Dataset Card for Charades

## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
  - [Dataset Summary](#dataset-summary)
  - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
  - [Languages](#languages)
- [Dataset Structure](#dataset-structure)
  - [Data Instances](#data-instances)
  - [Data Fields](#data-fields)
  - [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
  - [Curation Rationale](#curation-rationale)
  - [Source Data](#source-data)
  - [Annotations](#annotations)
  - [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
  - [Social Impact of Dataset](#social-impact-of-dataset)
  - [Discussion of Biases](#discussion-of-biases)
  - [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
  - [Dataset Curators](#dataset-curators)
  - [Licensing Information](#licensing-information)
  - [Citation Information](#citation-information)
  - [Contributions](#contributions)

## Dataset Description

- **Homepage:** https://prior.allenai.org/projects/charades
- **Repository:** https://github.com/gsig/charades-algorithms
- **Paper:** https://arxiv.org/abs/1604.01753
- **Leaderboard:** https://paperswithcode.com/sota/action-classification-on-charades
- **Point of Contact:** mailto: [email protected] 

### Dataset Summary

Charades is dataset composed of 9848 videos of daily indoors activities collected through Amazon Mechanical Turk. 267 different users were presented with a sentence, that includes objects and actions from a fixed vocabulary, and they recorded a video acting out the sentence (like in a game of Charades). The dataset contains 66,500 temporal annotations for 157 action classes, 41,104 labels for 46 object classes, and 27,847 textual descriptions of the videos

### Supported Tasks and Leaderboards

- `multilabel-action-classification`: The goal of this task is to classify actions happening in a video. This is a multilabel classification. The leaderboard is available [here](https://paperswithcode.com/sota/action-classification-on-charades)


### Languages

The annotations in the dataset are in English.

## Dataset Structure

### Data Instances

```
{
  "video_id": "46GP8",
  "video": "/home/amanpreet_huggingface_co/.cache/huggingface/datasets/downloads/extracted/3f022da5305aaa189f09476dbf7d5e02f6fe12766b927c076707360d00deb44d/46GP8.mp4",
  "subject": "HR43",
  "scene": "Kitchen",
  "quality": 6,
  "relevance": 7,
  "verified": "Yes",
  "script": "A person cooking on a stove while watching something out a window.",
  "objects": ["food", "stove", "window"],
  "descriptions": [
    "A person cooks food on a stove before looking out of a window."
  ],
  "labels": [92, 147],
  "action_timings": [
    [11.899999618530273, 21.200000762939453],
    [0.0, 12.600000381469727]
  ],
  "length": 24.829999923706055
}
```

### Data Fields

- `video_id`: `str` Unique identifier for each video.
- `video`: `str` Path to the video file
- `subject`: `str` Unique identifier for each subject in the dataset
- `scene`: `str` One of 15 indoor scenes in the dataset, such as Kitchen
- `quality`: `int` The quality of the video judged by an annotator (7-point scale, 7=high quality), -100 if missing
- `relevance`: `int` The relevance of the video to the script judged by an annotated (7-point scale, 7=very relevant), -100 if missing
- `verified`: `str` 'Yes' if an annotator successfully verified that the video matches the script, else 'No'
- `script`: `str` The human-generated script used to generate the video
- `descriptions`: `List[str]` List of descriptions by annotators watching the video
- `labels`: `List[int]` Multi-label actions found in the video. Indices from 0 to 156.
- `action_timings`: `List[Tuple[int, int]]` Timing where each of the above actions happened.
- `length`: `float` The length of the video in seconds

<details>
  <summary>
  Click here to see the full list of Charades class labels mapping:
  </summary>
  
  |id|Class|
  |--|-----|
  |c000 | Holding some clothes |
  |c001 | Putting clothes somewhere |
  |c002 | Taking some clothes from somewhere |
  |c003 | Throwing clothes somewhere |
  |c004 | Tidying some clothes |
  |c005 | Washing some clothes |
  |c006 | Closing a door |
  |c007 | Fixing a door |
  |c008 | Opening a door |
  |c009 | Putting something on a table |
  |c010 | Sitting on a table |
  |c011 | Sitting at a table |
  |c012 | Tidying up a table |
  |c013 | Washing a table |
  |c014 | Working at a table |
  |c015 | Holding a phone/camera |
  |c016 | Playing with a phone/camera |
  |c017 | Putting a phone/camera somewhere |
  |c018 | Taking a phone/camera from somewhere |
  |c019 | Talking on a phone/camera |
  |c020 | Holding a bag |
  |c021 | Opening a bag |
  |c022 | Putting a bag somewhere |
  |c023 | Taking a bag from somewhere |
  |c024 | Throwing a bag somewhere |
  |c025 | Closing a book |
  |c026 | Holding a book |
  |c027 | Opening a book |
  |c028 | Putting a book somewhere |
  |c029 | Smiling at a book |
  |c030 | Taking a book from somewhere |
  |c031 | Throwing a book somewhere |
  |c032 | Watching/Reading/Looking at a book |
  |c033 | Holding a towel/s |
  |c034 | Putting a towel/s somewhere |
  |c035 | Taking a towel/s from somewhere |
  |c036 | Throwing a towel/s somewhere |
  |c037 | Tidying up a towel/s |
  |c038 | Washing something with a towel |
  |c039 | Closing a box |
  |c040 | Holding a box |
  |c041 | Opening a box |
  |c042 | Putting a box somewhere |
  |c043 | Taking a box from somewhere |
  |c044 | Taking something from a box |
  |c045 | Throwing a box somewhere |
  |c046 | Closing a laptop |
  |c047 | Holding a laptop |
  |c048 | Opening a laptop |
  |c049 | Putting a laptop somewhere |
  |c050 | Taking a laptop from somewhere |
  |c051 | Watching a laptop or something on a laptop |
  |c052 | Working/Playing on a laptop |
  |c053 | Holding a shoe/shoes |
  |c054 | Putting shoes somewhere |
  |c055 | Putting on shoe/shoes |
  |c056 | Taking shoes from somewhere |
  |c057 | Taking off some shoes |
  |c058 | Throwing shoes somewhere |
  |c059 | Sitting in a chair |
  |c060 | Standing on a chair |
  |c061 | Holding some food |
  |c062 | Putting some food somewhere |
  |c063 | Taking food from somewhere |
  |c064 | Throwing food somewhere |
  |c065 | Eating a sandwich |
  |c066 | Making a sandwich |
  |c067 | Holding a sandwich |
  |c068 | Putting a sandwich somewhere |
  |c069 | Taking a sandwich from somewhere |
  |c070 | Holding a blanket |
  |c071 | Putting a blanket somewhere |
  |c072 | Snuggling with a blanket |
  |c073 | Taking a blanket from somewhere |
  |c074 | Throwing a blanket somewhere |
  |c075 | Tidying up a blanket/s |
  |c076 | Holding a pillow |
  |c077 | Putting a pillow somewhere |
  |c078 | Snuggling with a pillow |
  |c079 | Taking a pillow from somewhere |
  |c080 | Throwing a pillow somewhere |
  |c081 | Putting something on a shelf |
  |c082 | Tidying a shelf or something on a shelf |
  |c083 | Reaching for and grabbing a picture |
  |c084 | Holding a picture |
  |c085 | Laughing at a picture |
  |c086 | Putting a picture somewhere |
  |c087 | Taking a picture of something |
  |c088 | Watching/looking at a picture |
  |c089 | Closing a window |
  |c090 | Opening a window |
  |c091 | Washing a window |
  |c092 | Watching/Looking outside of a window |
  |c093 | Holding a mirror |
  |c094 | Smiling in a mirror |
  |c095 | Washing a mirror |
  |c096 | Watching something/someone/themselves in a mirror |
  |c097 | Walking through a doorway |
  |c098 | Holding a broom |
  |c099 | Putting a broom somewhere |
  |c100 | Taking a broom from somewhere |
  |c101 | Throwing a broom somewhere |
  |c102 | Tidying up with a broom |
  |c103 | Fixing a light |
  |c104 | Turning on a light |
  |c105 | Turning off a light |
  |c106 | Drinking from a cup/glass/bottle |
  |c107 | Holding a cup/glass/bottle of something |
  |c108 | Pouring something into a cup/glass/bottle |
  |c109 | Putting a cup/glass/bottle somewhere |
  |c110 | Taking a cup/glass/bottle from somewhere |
  |c111 | Washing a cup/glass/bottle |
  |c112 | Closing a closet/cabinet |
  |c113 | Opening a closet/cabinet |
  |c114 | Tidying up a closet/cabinet |
  |c115 | Someone is holding a paper/notebook |
  |c116 | Putting their paper/notebook somewhere |
  |c117 | Taking paper/notebook from somewhere |
  |c118 | Holding a dish |
  |c119 | Putting a dish/es somewhere |
  |c120 | Taking a dish/es from somewhere |
  |c121 | Wash a dish/dishes |
  |c122 | Lying on a sofa/couch |
  |c123 | Sitting on sofa/couch |
  |c124 | Lying on the floor |
  |c125 | Sitting on the floor |
  |c126 | Throwing something on the floor |
  |c127 | Tidying something on the floor |
  |c128 | Holding some medicine |
  |c129 | Taking/consuming some medicine |
  |c130 | Putting groceries somewhere |
  |c131 | Laughing at television |
  |c132 | Watching television |
  |c133 | Someone is awakening in bed |
  |c134 | Lying on a bed |
  |c135 | Sitting in a bed |
  |c136 | Fixing a vacuum |
  |c137 | Holding a vacuum |
  |c138 | Taking a vacuum from somewhere |
  |c139 | Washing their hands |
  |c140 | Fixing a doorknob |
  |c141 | Grasping onto a doorknob |
  |c142 | Closing a refrigerator |
  |c143 | Opening a refrigerator |
  |c144 | Fixing their hair |
  |c145 | Working on paper/notebook |
  |c146 | Someone is awakening somewhere |
  |c147 | Someone is cooking something |
  |c148 | Someone is dressing |
  |c149 | Someone is laughing |
  |c150 | Someone is running somewhere |
  |c151 | Someone is going from standing to sitting |
  |c152 | Someone is smiling |
  |c153 | Someone is sneezing |
  |c154 | Someone is standing up from somewhere |
  |c155 | Someone is undressing |
  |c156 | Someone is eating something |
</details>

### Data Splits


|             |train  |validation| test  |
|-------------|------:|---------:|------:|
|# of examples|1281167|50000     |100000 |


## Dataset Creation

### Curation Rationale

> Computer vision has a great potential to help our daily lives by searching for lost keys, watering flowers or reminding us to take a pill. To succeed with such tasks, computer vision methods need to be trained from real and diverse examples of our daily dynamic scenes. While most of such scenes are not particularly exciting, they typically do not appear on YouTube, in movies or TV broadcasts. So how do we collect sufficiently many diverse but boring samples representing our lives? We propose a novel Hollywood in Homes approach to collect such data. Instead of shooting videos in the lab, we ensure diversity by distributing and crowdsourcing the whole process of video creation from script writing to video recording and annotation.

### Source Data

#### Initial Data Collection and Normalization

> Similar to filming, we have a three-step process for generating a video. The first step is generating the script of the indoor video. The key here is to allow workers to generate diverse scripts yet ensure that we have enough data for each category. The second step in the process is to use the script and ask workers to record a video of that sentence being acted out. In the final step, we ask the workers to verify if the recorded video corresponds to script, followed by an annotation procedure.

#### Who are the source language producers?

Amazon Mechnical Turk annotators

### Annotations

#### Annotation process

> Similar to filming, we have a three-step process for generating a video. The first step is generating the script of the indoor video. The key here is to allow workers to generate diverse scripts yet ensure that we have enough data for each category. The second step in the process is to use the script and ask workers to record a video of that sentence being acted out. In the final step, we ask the workers to verify if the recorded video corresponds to script, followed by an annotation procedure.

#### Who are the annotators?

Amazon Mechnical Turk annotators

### Personal and Sensitive Information

Nothing specifically mentioned in the paper.

## Considerations for Using the Data

### Social Impact of Dataset

[More Information Needed]

### Discussion of Biases

[More Information Needed]

### Other Known Limitations

[More Information Needed]

## Additional Information

### Dataset Curators

AMT annotators

### Licensing Information

License for Non-Commercial Use

If this software is redistributed, this license must be included. The term software includes any source files, documentation, executables, models, and data.

This software and data is available for general use by academic or non-profit, or government-sponsored researchers.  It may also be used for evaluation purposes elsewhere.  This license does not grant the right to use this software or any derivation of it in a for-profit enterprise.  For commercial use, please contact The Allen Institute for Artificial Intelligence.

This license does not grant the right to modify and publicly release the data in any form.

This license does not grant the right to distribute the data to a third party in any form.

The subjects in this data should be treated with respect and dignity. This license only grants the right to publish short segments or still images in an academic publication where necessary to present examples, experimental results, or observations.

This software comes with no warranty or guarantee of any kind.  By using this software, the user accepts full liability.

The Allen Institute for Artificial Intelligence (C) 2016.

### Citation Information

```bibtex
@article{sigurdsson2016hollywood,
    author = {Gunnar A. Sigurdsson and G{\"u}l Varol and Xiaolong Wang and Ivan Laptev and Ali Farhadi and Abhinav Gupta},
    title = {Hollywood in Homes: Crowdsourcing Data Collection for Activity Understanding},
    journal = {ArXiv e-prints},
    eprint = {1604.01753}, 
    year = {2016},
    url = {http://arxiv.org/abs/1604.01753},
}
```

### Contributions

Thanks to [@apsdehal](https://github.com/apsdehal) for adding this dataset.