--- dataset_info: features: - name: id dtype: string - name: conversations list: - name: role dtype: string - name: content dtype: string splits: - name: train num_bytes: 277884785 num_examples: 160000 download_size: 126665150 dataset_size: 277884785 configs: - config_name: default data_files: - split: train path: data/train-* ---
ð Homepage ⢠ð Paper ⢠ð€ Data (PVD-160k) ⢠ð€ Model (PVD-160k-Mistral-7b) ⢠ð» Code
We observe that current *large multimodal models (LMMs)* still struggle with seemingly straightforward reasoning tasks that require precise perception of low-level visual details, such as identifying spatial relations or solving simple mazes. In particular, this failure mode persists in question-answering tasks about vector graphicsâimages composed purely of 2D objects and shapes. ![Teaser](https://github.com/MikeWangWZHL/VDLM/blob/main/figures/teaser.png?raw=true) To solve this challenge, we propose **Visually Descriptive Language Model (VDLM)**, a visual reasoning framework that operates with intermediate text-based visual descriptionsâSVG representations and learned Primal Visual Description, which can be directly integrated into existing LLMs and LMMs. We demonstrate that VDLM outperforms state-of-the-art large multimodal models, such as GPT-4V, across various multimodal reasoning tasks involving vector graphics. See our [paper](https://arxiv.org/abs/2404.06479) for more details. ![Overview](https://github.com/MikeWangWZHL/VDLM/blob/main/figures/overview.png?raw=true)