# Download datasets: | |
* Download and decompress tsv file from here: https://github.com/google-research-datasets/wit/blob/main/DATA.md | |
* Use `prepare_wit.py` to download images from Wikipedia as annotated on each TSV file. | |
* Use `scale_converter.py` to remove corrupt images and resize suitable images to 224x224 | |
* Use `join_datasets_custom_split.py` to group all JSONs from different subsets of the dataset together | |
* Use `discard_incorrect_files.py` to filter out images that we were not able to convert. | |
* Finally, use `run-clip.sh` to train. |