--- license: cc-by-4.0 task_categories: - video-text-to-text - visual-question-answering - image-to-text language: - en size_categories: - 10M The OmniCorpus contains three sections: - **OmniCorpus-CC**: processed from dumps in Common Crawl from 2013 to Nov./Dec. 2023. - **OmniCorpus-CW**: sourced from Chinese internet resources, will be availiable on [OpenDataLab](https://opendatalab.com/) platform. - **OmniCorpus-YT**: samples Youtube video frames as images and collects subtitles as texts. Code for pre-training, evaluating, main body extracting, and filtering have been released in the official [repository](https://github.com/OpenGVLab/OmniCorpus). A pre-trained model is availiable [here](https://huggingface.co/Qingyun/OmniCorpus-InternVL). # Usages The image-text interleaved documents are recommanded for the following usages: - Pre-training multimodal large language model (MLLM): Recent MLLMs (such as Flamingo series, EMU series, IDEFICS series, MM1, Cambrian-1, and xGen-MM) have shown that image-text interleaved data aids multimodal in-context learning and maintains the capabilities of large language models during multimodal fine-tuning. - Long text-image retrieval: We provide image-text similarities calculated with CLIP, which can convert the documents to image-text retrieval dataset with longer text. A retrieval model pre-trained on such data can retrieval images based on longer text, which can be used for multimodal RAG, converting pure text to multimodal sample, etc. - Source for futher dataset research: Our data is large-scale, which can serve as the source for researches for data curation strategies. We provide many useful attributes as metadata for each document, which can enrich the filtering strategy and reduce the cost. - ...... # Data Format Following common practices, the data is organized into Parquet file format. You might encounter errors when using `pandas.read_parquet` (because the data structure contains nested elements). We recommend using fastparquet to load the parquet files. ```Python import fastparquet df = fastparquet.ParquetFile(parquet_file_path).to_pandas() # You can also use iter_batches parquet_file = pq.ParquetFile(filepath) for batch in parquet_file.iter_batches(): df = batch.to_pandas() ``` You can convert the i-th document and convert it into a dictionary. ```Python doc_dict = df.iloc[i].to_dict() ``` The document format is as follow: ```json { 'id': , 'images': , 'texts': } ``` the images and texts can be loaded with `lambda s: json.loads(s)` ```json 'images': [ , None, , None, ], 'texts': [ None, None, , ] ``` The frame can be sampled from downloaded Youtube videos, we provide a python sampling tool: ```python import os import sys import yt_dlp # pip install yt-dlp import ffmpeg # brew install ffmpeg; pip install ffmpeg-python import traceback from multiprocessing import Pool def download_hls_url(youtube_id): video_url = f"https://www.youtube.com/watch?v={youtube_id}" ydl_opts = { 'format': 'best', 'noplaylist': True, 'quiet': True, } with yt_dlp.YoutubeDL(ydl_opts) as ydl: info = ydl.extract_info(video_url, download=False) return info['url'] def extract_frame(hls_url, timestamp, output_file): try: ( ffmpeg .input(hls_url, ss=timestamp, protocol_whitelist='file,http,https,tcp,tls,httpproxy') .output(output_file, vframes=1) .run(quiet=True, capture_stdout=True, capture_stderr=True) ) except ffmpeg.Error as e: print(f"Error extracting frame at timestamp {timestamp}: {e}") print("FFmpeg stderr output:\n", e.stderr.decode()) traceback.print_exc() def extract_frames_with_hls(youtube_id, timestamps, output_dir='frames'): if not os.path.exists(output_dir): os.makedirs(output_dir) hls_url = download_hls_url(youtube_id) tasks = [(hls_url, timestamp, os.path.join(output_dir, f"{timestamp}.jpg")) for timestamp in timestamps] with Pool() as pool: pool.starmap(extract_frame, tasks) if __name__ == "__main__": extract_frames_with_hls("1xGiPUeevCM", [19.000000, 23.000000, 28.000000, 32.000000, 45.000000, 54.000000, 57.000000, 67.000000]) ``` # License OmniCorpus is released under a [CC-BY-4.0](https://creativecommons.org/licenses/by/4.0/deed.en) license, with the primary intent of supporting research activities. # Citation ``` @article{li2024omnicorpus, title={OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text}, author={Li, Qingyun and Chen, Zhe and Wang, Weiyun and Wang, Wenhai and Ye, Shenglong and Jin, Zhenjiang and others}, journal={arXiv preprint arXiv:2406.08418}, year={2024} } ```