Update README.md
Browse files
README.md
CHANGED
@@ -11,6 +11,9 @@ datasets:
|
|
11 |
---
|
12 |
|
13 |
## MiniCPM-V 2.0
|
|
|
|
|
|
|
14 |
**MiniCPM-V 2.8B** is a strong multimodal large language model for efficient end-side deployment. The model is built based on SigLip-400M and [MiniCPM-2.4B](https://github.com/OpenBMB/MiniCPM/), connected by a perceiver resampler. Our latest version, **MiniCPM-V 2.0** has several notable features.
|
15 |
|
16 |
- 🔥 **State-of-the-art Performance.**
|
@@ -132,4 +135,33 @@ Please see the info about MiniCPM-V 1.0 [here](https://huggingface.co/openbmb/Mi
|
|
132 |
|
133 |
#### Statement
|
134 |
* As a LLM, MiniCPM-V 2.0 generates contents by learning a large mount of texts, but it cannot comprehend, express personal opinions or make value judgement. Anything generated by MiniCPM-V 2.0 does not represent the views and positions of the model developers
|
135 |
-
* We will not be liable for any problems arising from the use of the MinCPM-V open Source model, including but not limited to data security issues, risk of public opinion, or any risks and problems arising from the misdirection, misuse, dissemination or misuse of the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
---
|
12 |
|
13 |
## MiniCPM-V 2.0
|
14 |
+
|
15 |
+
[GitHub](https://github.com/OpenBMB/MiniCPM-V) | [Demo](http://120.92.209.146/)
|
16 |
+
|
17 |
**MiniCPM-V 2.8B** is a strong multimodal large language model for efficient end-side deployment. The model is built based on SigLip-400M and [MiniCPM-2.4B](https://github.com/OpenBMB/MiniCPM/), connected by a perceiver resampler. Our latest version, **MiniCPM-V 2.0** has several notable features.
|
18 |
|
19 |
- 🔥 **State-of-the-art Performance.**
|
|
|
135 |
|
136 |
#### Statement
|
137 |
* As a LLM, MiniCPM-V 2.0 generates contents by learning a large mount of texts, but it cannot comprehend, express personal opinions or make value judgement. Anything generated by MiniCPM-V 2.0 does not represent the views and positions of the model developers
|
138 |
+
* We will not be liable for any problems arising from the use of the MinCPM-V open Source model, including but not limited to data security issues, risk of public opinion, or any risks and problems arising from the misdirection, misuse, dissemination or misuse of the model.
|
139 |
+
|
140 |
+
## Multimodal Projects of Our Team
|
141 |
+
|
142 |
+
[VisCPM](https://github.com/OpenBMB/VisCPM/tree/main) | [RLHF-V](https://github.com/RLHF-V/RLHF-V) | [LLaVA-UHD](https://github.com/thunlp/LLaVA-UHD) | [Muffin](https://github.com/thunlp/Muffin/tree/main)
|
143 |
+
|
144 |
+
## Citation
|
145 |
+
|
146 |
+
If you find our work helpful, please consider citing the following papers
|
147 |
+
|
148 |
+
```bib
|
149 |
+
@article{yu2023rlhf,
|
150 |
+
title={Rlhf-v: Towards trustworthy mllms via behavior alignment from fine-grained correctional human feedback},
|
151 |
+
author={Yu, Tianyu and Yao, Yuan and Zhang, Haoye and He, Taiwen and Han, Yifeng and Cui, Ganqu and Hu, Jinyi and Liu, Zhiyuan and Zheng, Hai-Tao and Sun, Maosong and others},
|
152 |
+
journal={arXiv preprint arXiv:2312.00849},
|
153 |
+
year={2023}
|
154 |
+
}
|
155 |
+
@article{viscpm,
|
156 |
+
title={Large Multilingual Models Pivot Zero-Shot Multimodal Learning across Languages},
|
157 |
+
author={Jinyi Hu and Yuan Yao and Chongyi Wang and Shan Wang and Yinxu Pan and Qianyu Chen and Tianyu Yu and Hanghao Wu and Yue Zhao and Haoye Zhang and Xu Han and Yankai Lin and Jiao Xue and Dahai Li and Zhiyuan Liu and Maosong Sun},
|
158 |
+
journal={arXiv preprint arXiv:2308.12038},
|
159 |
+
year={2023}
|
160 |
+
}
|
161 |
+
@article{xu2024llava-uhd,
|
162 |
+
title={{LLaVA-UHD}: an LMM Perceiving Any Aspect Ratio and High-Resolution Images},
|
163 |
+
author={Xu, Ruyi and Yao, Yuan and Guo, Zonghao and Cui, Junbo and Ni, Zanlin and Ge, Chunjiang and Chua, Tat-Seng and Liu, Zhiyuan and Huang, Gao},
|
164 |
+
journal={arXiv preprint arXiv:2403.11703},
|
165 |
+
year={2024}
|
166 |
+
}
|
167 |
+
```
|