---
license: cc-by-nc-4.0
language:
- en
pipeline_tag: image-text-to-text
tags:
- nvidia
- NVLM
- pytorch
- multimodal
- conversational
library_name: transformers
---
\nPlease describe the image shortly.'
response = model.chat(tokenizer, pixel_values, question, generation_config)
print(f'User: {question}\nAssistant: {response}')
```
## Software Integration
**Runtime Engine(s)**
* PyTorch
**Supported Hardware Microarchitecture Compatibility:**
* NVIDIA Hopper
**[Preferred/Supported] Operating System(s):**
* Linux
## Inference
**Engine:** PyTorch
**Test Hardware:**
* H100
## Model Version(s)
* v1.0-D (NVLM-D)
## Training, Testing, and Evaluation Datasets
### Pre-Training Dataset
**Link**
* [See Table 4](https://arxiv.org/abs/2409.11402)
**Data Collection Method by dataset**
* Hybrid: Automated, Human, Synthetic, Unknown
**Labeling Method by dataset**
* Hybrid: Automated, Human, Synthetic, Unknown
**Properties**
* Trained on image captions, image-text pairs, natural images, charts, documents, scene descriptions, and mathematical reasoning.
### Supervised Fine-Tuning Dataset
**Link**
* [See Table 6](https://arxiv.org/abs/2409.11402)
**Data Collection Method by dataset**
* Hybrid: Automated, Human, Synthetic, Unknown
**Labeling Method by dataset**
* Hybrid: Automated, Human, Synthetic, Unknown
**Properties**
* Trained on image captions; general knowledge; image-text pairs; natural images; charts; diagrams; documents; scene descriptions; science diagrams, lessons, textbook data, and question-answer pairs; visual instruction tuning; and mathematical reasoning.
### Evaluation Dataset
**Link**
* [See Section 6.1, "Benchmark"](https://arxiv.org/abs/2409.11402)
**Data collection method by dataset**
* Human
**Labeling method by dataset**
* Human
**Properties**
* Evaluated on general knowledge, visual answering, chart understanding, table, optical character recognition, and mathematical reasoning.
## Correspondence to
Wenliang Dai* (wdai@nvidia.com), Nayeon Lee* (nayeonl@nvidia.com), Boxin Wang* (boxinw@nvidia.com), Zhuolin Yang* (zhuoliny@nvidia.com), Wei Ping* (wping@nvidia.com)
*Equal contribution
## Citation
@article{nvlm2024,
title={NVLM: Open Frontier-Class Multimodal LLMs},
author={Dai, Wenliang and Lee, Nayeon and Wang, Boxin and Yang, Zhuolin and Liu, Zihan and Barker, Jon and Rintamaki, Tuomas and Shoeybi, Mohammad and Catanzaro, Bryan and Ping, Wei},
journal={arXiv preprint},
year={2024}}
## Ethical Considerations
NVIDIA believes Trustworthy AI is a shared responsibility and we have established policies and practices to enable development for a wide array of AI applications. When downloaded or used in accordance with our terms of service, developers should work with their internal model team to ensure this model meets requirements for the relevant industry and use case and addresses unforeseen product misuse.
Please report security vulnerabilities or NVIDIA AI Concerns [here](https://www.nvidia.com/en-us/support/submit-security-vulnerability/).