Datasets:

Modalities:
Image
Text
Formats:
json
ArXiv:
Libraries:
Datasets
Dask
License:
davidanugraha commited on
Commit
e2c8e50
1 Parent(s): 240c08a

Update README.md but still need to describe how to do evaluation

Browse files
Files changed (1) hide show
  1. README.md +121 -0
README.md CHANGED
@@ -57,3 +57,124 @@ configs:
57
  - split: train
58
  path: hf_prompt/train_task2/*
59
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  - split: train
58
  path: hf_prompt/train_task2/*
59
  ---
60
+
61
+ # WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines
62
+
63
+ WorldCuisines is a massive-scale visual question answering (VQA) benchmark for multilingual and multicultural understanding through global cuisines. The dataset contains text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark as of 17 October 2024.
64
+
65
+ ## Overview
66
+ We develop both a VQA dataset (**WC-VQA**) and a curated Knowledge Base (KB) for world cuisines (**WC-KB**). The **WC-VQA** dataset is constructed using **WC-KB**, which serves as the primary data source. We design two tasks as follows:
67
+
68
+ #### Task 1: Dish Name Prediction
69
+ This task requires predicting the name of a dish based on its image, a question, and contextual information. It is divided into three subtasks, each with a distinct query type:
70
+ - (a) **No-context question**: Predict the dish name without additional context.
71
+ - (b) **Contextualized question**: Predict the dish name with provided contextual information.
72
+ - (c) **Adversarial contextualized question**: Predict the dish name with misleading or adversarial context.
73
+ #### Task 2: Location Prediction
74
+ This task involves predicting the location where the dish is commonly consumed or originated, based on the dish image, a question, and contextual information.
75
+
76
+ **WC-KB** encompasses 2,414 dishes worldwide, including 6,045 images and metadata, covering both coarse-grained (e.g., stew) and fine-grained categories (e.g., beef stew), locations, and regional cuisines. It also features multilingual translations of 90 crowd-sourced prompt templates and 401 parallel data entries (i.e., multilingual information) for location and regional cuisine information.
77
+ From **WC-KB**, we construct **WC-VQA**, a multilingual parallel VQA dataset with 1 million samples encompassing over 30 languages and dialects, including various varieties and registers, such as formal and casual styles, with high-quality human annotations. The VQA is designed to evaluate models' ability to understand cultural food names and their origins.
78
+
79
+ We provide **WC-VQA** evaluation datasets in two sizes (12,000 and 60,000 instances) alongside a training dataset (1,080,000 instances).
80
+ The table below provides more detailed statistics regarding the number of VQA instances and images for each data split.
81
+
82
+ | Data Split | Task 1 (Dish Name) | Task 2 (Location) | Total VQA |
83
+ |---------------------|--------------------------------------------------------------|-------------------|-----------|
84
+ | | (a) no-context | (b) contextualized | (c) adversarial | | |
85
+ | | # VQA | # Images | # VQA | # Images | # VQA | # Images | # VQA | # Images | |
86
+ | **Train (1M)** | 270,300 | 3,383 | 267,930 | 3,555 | 271,770 | 3,589 | 270,000 | 3,361 | 1,080,000 |
87
+ | **Test Small (12k)** | 3,000 | 100 | 3,000 | 100 | 3,000 | 100 | 3,000 | 100 | 12,000 |
88
+ | **Test Large (60k)** | 15,000 | 500 | 15,000 | 500 | 15,000 | 499 | 15,000 | 499 | 60,000 |
89
+
90
+ ## Dataset Construction
91
+
92
+ Our data sources are gathered from [Wikipedia](https://wikipedia.org) and [Wikimedia Commons](https://commons.wikimedia.org) to ensure they can be easily redistributed under an accepted open-source license. The data construction process involves four key steps:
93
+ 1. Dish selection
94
+ 2. Metadata annotation
95
+ 3. Quality assurance
96
+ 4. Data compilation.
97
+
98
+ ### Dish Selection
99
+
100
+ We compile a comprehensive list of dish names sourced from Wikipedia. We manually review pages that feature lists of dishes to determine whether each dish is a specialty unique to a specific culture, as we aim to focus on dishes that have distinct cultural significance. We exclude generic categories, such as ice cream, which lacks a specific cultural association. We ensure that each dish on our list has its own dedicated Wikipedia page. If a dish does not have a Wikipedia page, it is also excluded from our compilation. This meticulous approach ensures that our dataset is both culturally relevant and well-documented.
101
+
102
+ ### Metadata Annotation
103
+
104
+ Given a dish name and its corresponding Wikipedia page link, annotators manually compile metadata based on the provided information. This metadata includes:
105
+ - **Visual Representation**: Images sourced from Wikimedia Commons are included, along with their license information.
106
+ - **Categorization**: Dishes are classified into both coarse-grained (e.g., rice, bread) and fine-grained (e.g., fried rice, flatbread) categories.
107
+ - **Description**: Annotators provide a description of each dish based on the content from its Wikipedia page, avoiding the use of the dish's name, origin, or any distinctive keywords that uniquely identify the dish.
108
+ - **Cuisine**: The dish's origin cuisine and any cuisines with which it is strongly associated.
109
+ - **Geographic Distribution**: This includes the dish's associated countries, area (city or region), and broader continental region.
110
+
111
+ ### Quality Assurance
112
+
113
+ Before beginning the quality assurance process, we first identify common issues that arise during annotation and develop automated rules to detect easily identifiable annotation errors, such as incorrect string formatting. Annotators are then asked to correct these errors. To further ensure data quality and validity, we conduct several rounds of quality assurance:
114
+
115
+ 1. **Image Quality**: We remove instances where images are blurry, dark, or contain distracting elements such as people or other dishes. We also verify image licenses by cross-referencing them with information on Wikimedia Commons.
116
+ 2. **Categorization and Descriptions**: We refine dish categorization and descriptions, ensuring consistency in category assignments and keeping descriptions free from "information breaches" (e.g., excluding regional details from the description).
117
+ 3. **Cuisine Names and Geographic Accuracy**: We standardize cuisine names and meticulously review all country and area information for accuracy.
118
+
119
+ This comprehensive approach guarantees the integrity and reliability of our dataset.
120
+
121
+ ### Data Compilation
122
+
123
+ In this phase, we verify the overall quality check done by annotators, identifying any potential inconsistencies missed during quality assurance. We then compile the dataset by collecting the metadata into a single file.
124
+
125
+ ## VQA Generation
126
+
127
+ In this phase, we generate VQA data by sampling from **WC-KB**. A VQA data entry comprises a visual image, question text, and answer text. This process involves four stages:
128
+
129
+ 1. Conducting a similarity search for dish names
130
+ 2. Constructing questions and contexts
131
+ 3. Translating these elements into multiple languages
132
+ 4. Generating the VQA triplets.
133
+
134
+ ### Dish Names Similarity Search
135
+
136
+ To identify similar dishes in our dataset, we follow the approach from Winata et al. (2024) to employ a multilingual model E5$_\text{LARGE}$ Instruct (Wang et al. 2024) for computing text embeddings. Formally, given a dish $x$ with name $x_{\text{name}}$ and text description $x_{\text{desc}}$, we use a multilingual model $\theta$ to compute the embedding vector $v_x = \theta(\{x_\text{name};x_\text{desc}\})$, and then apply cosine similarity to compute a score $s = \text{similarity}(v_i, v_j)$ between dish $i$ and dish $j$. For each dish, we consider the top-$k$ most similar dishes to generate distractors in the multiple choice questions.
137
+
138
+ ### Question and Context Construction
139
+
140
+ Dish name prediction (Task 1) is divided into three question variations depending on the context:
141
+
142
+ 1. **No-context question**: Simply asks for the name of the dish without any provided context.
143
+ 2. **Contextualized question**: Provides additional information related to cuisine or location.
144
+ 3. **Adversarial contextualized question**: Similar to contextualized questions, but may include misleading location information to assess the model's robustness to irrelevant details.
145
+
146
+ For regional cuisine prediction (Task 2), only a basic question without any provided context is available.
147
+
148
+ ### Multiple Language Translation
149
+
150
+ #### Question and Context
151
+
152
+ All questions and contexts are initially collected in English and are then carefully translated by native speakers into 30 language varieties: 23 different languages, with 7 languages having two varieties. Translators prioritize naturalness, followed by diversity in translations when duplication occurs.
153
+
154
+ #### Food Name Alias
155
+
156
+ Using Wikipedia pages as our primary source, we verify if the English page has translations available in other languages. This enables us to extract dish names in multiple languages and compile them as translations for each dish. We utilize both the Wikipedia page titles in various languages and alias text from the English page. These translations enhance cultural relevance and accuracy for multilingual prompts. When translation is unavailable, we use the English name as the default.
157
+
158
+ #### Locations and Cuisines
159
+
160
+ With over 400 unique locations, including countries, cities, and areas, we first translate the English locations into other languages using GPT-4, followed by proofreading by native speakers. Regional cuisine names (the adjective form of the location in English) are translated in the same manner.
161
+
162
+ #### Morphological Inflections
163
+
164
+ In languages with rich inflectional morphology (e.g., Czech or Spanish), words are modified to express different grammatical categories (e.g., number, gender, or case). We provide a framework for human translators to use natural inflections in the prompt template while keeping the inflections as few as possible.
165
+
166
+ ### Generating VQA Triplets
167
+
168
+ To ensure no overlap between training and testing subsets, we split the dishes and multilingual questions into two subsets. For each subset, we randomly sample dish and question pairs. We use the dish entry in our KB dataset to select the image and inject the location into the context, if applicable. Answer candidates for multiple-choice questions are picked using the similarity search process. We repeat this process until the desired number of training or test samples is reached, discarding duplicates.
169
+
170
+ ## Ethical Considerations
171
+
172
+ Our research focuses on evaluating VLMs within the context of multilingual and multicultural VQA, a field that holds significant implications for diverse multilingual communities. We are committed to conducting our data collection and evaluations with the highest standards of transparency and fairness. To achieve this, we have adopted a crowd-sourcing approach for the annotation process, inviting volunteers to contribute and become co-authors if they provide significant contributions. We follow the guidelines from ACL for authorship eligibility as shown in https://www.aclweb.org/adminwiki/index.php/Authorship_Changes_Policy_for_ACL_Conference_Papers. In line with our commitment to openness and collaboration, we will release our dataset under an open-source license, CC-BY-SA 4.0.
173
+
174
+ ## Contact
175
+
176
+ E-mail: [Genta Indra Winata]([email protected]) and [Frederikus Hudi]([email protected])
177
+
178
+ ## Citation
179
+
180
+ If you find this dataset useful, please cite the following works