clip-spanish / README.md
edugp's picture
Add all necessary files to replicate training run
2daf3c7
|
raw
history blame
551 Bytes
# Download datasets:
* Download and decompress tsv file from here: https://github.com/google-research-datasets/wit/blob/main/DATA.md
* Use `prepare_wit.py` to download images from Wikipedia as annotated on each TSV file.
* Use `scale_converter.py` to remove corrupt images and resize suitable images to 224x224
* Use `join_datasets_custom_split.py` to group all JSONs from different subsets of the dataset together
* Use `discard_incorrect_files.py` to filter out images that we were not able to convert.
* Finally, use `run-clip.sh` to train.