--- dataset_info: config_name: music features: - name: image dtype: image - name: download_url dtype: string - name: instance_name dtype: string - name: date dtype: string - name: additional_info dtype: string - name: date_scrapped dtype: string - name: compilation_info dtype: string - name: rendering_filters dtype: string - name: assets sequence: string - name: category dtype: string - name: uuid dtype: string - name: length dtype: string - name: difficulty dtype: string splits: - name: validation num_bytes: 25180255.0 num_examples: 300 download_size: 24944162 dataset_size: 25180255.0 configs: - config_name: music data_files: - split: validation path: music/validation-* --- # Image2Struct - Music Sheet [Paper](TODO) | [Website](https://crfm.stanford.edu/helm/image2structure/latest/) | Datasets ([Webpages](https://huggingface.co/datasets/stanford-crfm/i2s-webpage), [Latex](https://huggingface.co/datasets/stanford-crfm/i2s-latex), [Music sheets](https://huggingface.co/datasets/stanford-crfm/i2s-musicsheet)) | [Leaderboard](https://crfm.stanford.edu/helm/image2structure/latest/#/leaderboard) | [HELM repo](https://github.com/stanford-crfm/helm) | [Image2Struct repo](https://github.com/stanford-crfm/image2structure) **License:** [Apache License](http://www.apache.org/licenses/) Version 2.0, January 2004 ## Dataset description Image2struct is a benchmark for evaluating vision-language models in practical tasks of extracting structured information from images. This subdataset focuses on Music sheets. The model is given an image of the expected output with the prompt: ``` Please generate the Lilypond code to generate a music sheet that looks like this image as much as feasibly possible. This music sheet was created by me, and I would like to recreate it using Lilypond. ``` The data was collected from IMSLP and has no ground truth. This means that while we prompt models to output some [Lilypond](https://lilypond.org/) code to recreate the image of the music sheet, we do not have access to a Lilypond code that could reproduce the image and would act as a "ground-truth". There is no **wild** subset as this already constitutes a dataset without ground-truths. ## Uses To load the subset `music` of the dataset to be sent to the model under evaluation in Python: ```python import datasets datasets.load_dataset("stanford-crfm/i2s-musicsheet", "music", split="validation") ``` To evaluate a model on Image2Musicsheet (equation) using [HELM](https://github.com/stanford-crfm/helm/), run the following command-line commands: ```sh pip install crfm-helm helm-run --run-entries image2musicsheet,model=vlm --models-to-run google/gemini-pro-vision --suite my-suite-i2s --max-eval-instances 10 ``` You can also run the evaluation for only a specific `difficulty`: ```sh helm-run --run-entries image2musicsheet:difficulty=hard,model=vlm --models-to-run google/gemini-pro-vision --suite my-suite-i2s --max-eval-instances 10 ``` For more information on running Image2Struct using [HELM](https://github.com/stanford-crfm/helm/), refer to the [HELM documentation](https://crfm-helm.readthedocs.io/) and the article on [reproducing leaderboards](https://crfm-helm.readthedocs.io/en/latest/reproducing_leaderboards/). ## Citation **BibTeX:** ```tex @misc{roberts2024image2struct, title={Image2Struct: A Benchmark for Evaluating Vision-Language Models in Extracting Structured Information from Images}, author={Josselin Somerville Roberts and Tony Lee and Chi Heem Wong and Michihiro Yasunaga and Yifan Mai and Percy Liang}, year={2024}, eprint={TBD}, archivePrefix={arXiv}, primaryClass={TBD} } ```