The viewer is disabled because this dataset repo requires arbitrary Python code execution. Please consider
removing the
loading script
and relying on
automated data support
(you can use
convert_to_parquet
from the datasets
library). If this is not possible, please
open a discussion
for direct help.
Dataset Card for VASR
Dataset Description
VASR is a challenging dataset for evaluating computer vision commonsense reasoning abilities. Given a triplet of images, the task is to select an image candidate B' that completes the analogy (A to A' is like B to what?). Unlike previous work on visual analogy that focused on simple image transformations, we tackle complex analogies requiring understanding of scenes. Our experiments demonstrate that state-of-the-art models struggle with carefully chosen distractors (±53%, compared to 90% human accuracy).
- Homepage: https://vasr-dataset.github.io/
- Colab https://colab.research.google.com/drive/1HUg0aHonFDK3hVFrIRYdSEfpUJeY-4dI
- Repository: https://github.com/vasr-dataset/vasr/tree/main/experiments
- Paper: https://arxiv.org/abs/2212.04542
- Leaderboard: https://vasr-dataset.github.io/
- Point of Contact: [email protected]
Supported Tasks and Leaderboards
https://vasr.github.io/leaderboard. https://paperswithcode.com/dataset/vasr.
How to Submit Predictions?
To submit predictions, please send a prediction CSV file to [email protected] / [email protected].
The prediction file should include a "B'" column with the predicted candidate name that best solves the analogy, and an index from 1 to 4 indicating the location of the predicted candidate in the given candidate list.
An example prediction file is available HERE.
A submission is allowed once a week, and you will receive a response within a week.
Colab notebook code for VASR evaluation with ViT
https://colab.research.google.com/drive/1HUg0aHonFDK3hVFrIRYdSEfpUJeY-4dI
Languages
English.
Dataset Structure
Data Fields
A: datasets.Image() - the first input image, A:A'.
A': datasets.Image() - the second input image, different from A in a single key, A:A'.
B: datasets.Image() - the third input image, has the same different item as A, B:B'.
B': datasets.Image() - the forth image, which is the analogy solution. Different from B in a single key (the same different one as in A:A'), B:B'. Hidden in the test set.
candidates_images: [datasets.Image()] - a list of candidate images solutions to the analogy.
label: datasets.Value("int64") - the index of the ground-truth solution. Hidden in the test set.
candidates: [datasets.Value("string")] - a list of candidate string solutions to the analogy.
Data Splits
There are three splits, TRAIN, VALIDATION, and TEST.
Since there are four candidates and one solution, random chance is 25%.
Dataset Creation
We leverage situation recognition annotations and the CLIP model to generate a large set of 500k candidate analogies. There are two types of labels:
- Silver labels, obtained from the automatic generation.
- Gold labels, obtained from human annotations over the silver annotations.
In the huggingface version we provide only the gold labeled dataset. Please refer to the project website download page if you want to download the silver labels version.
Annotations
Annotation process
We paid Amazon Mechanical Turk Workers to solve analogies, five annotators for each analogy. Workers were asked to select the image that best solves the analogy. The resulting dataset is composed of the 3,820 instances agreed upon with a majority vote of at least 3 annotators, which was obtained in 93% of the cases.
Considerations for Using the Data
All associations were obtained with human annotators.
All used images are from the imSitu dataset (http://imsitu.org/)
Using this data is allowed for academic research alone.
Licensing Information
CC-By 4.0
Citation Information
NA
- Downloads last month
- 129