Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
Dask
License:
OBELICS / README.md
HugoLaurencon's picture
Update README.md
aff6f31
|
raw
history blame
3.63 kB
---
dataset_info:
features:
- name: images
sequence: string
- name: metadata
dtype: string
- name: general_metadata
dtype: string
- name: texts
sequence: string
splits:
- name: train
num_bytes: 715724717192
num_examples: 141047697
download_size: 71520629655
dataset_size: 715724717192
license: cc-by-4.0
language:
- en
pretty_name: OBELISC
size_categories:
- 100M<n<1B
---
# Dataset Card for OBELISC
## Dataset Description
- **Repository: https://github.com/huggingface/OBELISC**
- **Paper: OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents**
- **Point of Contact: [email protected]**
### Dataset Summary
`OBELISC` is an open, massive and curated collection of interleaved image-text web documents, containing 141M documents, 115B text tokens and 353M images.
This dataset can be used to train large multimodal models, significantly improving their reasoning abilities compared to models trained solely on image/text pairs. Please refer to our paper for further details about the construction of the dataset, quantitative and qualitative analyses of `OBELISC`, and experiments we conducted.
### Languages
English
## Data Fields
There are 4 fields: `images`, `texts`, `metadata` and `general_metadata`.
For each example, the data in the columns `images` and `texts` are two lists of the same size, where for each index, one element and only one is not `None`.
For example, for the web document `<image_1>text<image_2>`, in `images`, we have `[image_1,None,image_2]` and in `texts` we have `[None,text,None]`.
The images are replaced by their URLs, and the users have to download them themselves, for example with the library `img2dataset`.
In `metadata`, there is a string that can be transformed into a list with `json.loads(example["metadata"])`. This list will have the same size as the lists of images and texts, and will have a dictionary for each index where there is an image, and a `None` value when there is a text. This dictionary will contain the metadata of the image (original source document, unformatted source, alt-text if present, ...).
Finally, in `general_metadata`, there is a string that can be transformed into a dictionary, containing the URL of the document, and information about its location in the Common Crawl data.
## Data Splits
There is only one split, `train`, that contains 141,047,697 examples.
## Size
`OBELISC` with images replaced by their URLs weighs 666.6 GB (unwanted!) in arrow format and 377 GB in this uploaded `parquet` format.
### Visualization of OBELISC documents
https://huggingface.co/spaces/HuggingFaceM4/obelisc_visualization
### Research paper
https://arxiv.org/abs/2306.16527
### GitHub repository
https://github.com/huggingface/OBELISC
## Terms of Use
By using the dataset, you agree to comply with the original licenses of the source content as well as the dataset license (CC-BY-4.0). Additionally, if you use this dataset to train a Machine Learning model, you agree to disclose your use of the dataset when releasing the model or an ML application using the model.
### Licensing Information
License CC-BY-4.0.
### Citation Information
If you are using this dataset, please cite
```
@inproceedings{
lauren{\c{c}}on2023obe,
title={OBELISC: An Open Web-Scale Filtered Dataset of Interleaved Image-Text Documents},
author={Hugo Lauren{\c{c}}on and Lucile Saulnier and L{\'e}o Tronchon and Stas Bekman and Amanpreet Singh and Anton Lozhkov and Thomas Wang and Siddharth Karamcheti and Alexander M Rush and Douwe Kiela and Matthieu Cord and Victor Sanh},
year={2023}
}
```