BAAI
/

QuanSun commited on
Commit
f1fbfac
1 Parent(s): a48e1ea

update README

Browse files
Files changed (1) hide show
  1. README.md +2 -20
README.md CHANGED
@@ -3,28 +3,10 @@ language:
3
  - en
4
  ---
5
 
6
- <div align='center'>
7
- <h1>Generative Multimodal Models are In-Context Learners</h1h1>
8
- <h3><a href="">Generative Multimodal Models are In-Context Learners</a></h3>
9
 
10
- [Quan Sun](https://github.com/Quan-Sun)<sup>1*</sup>, [Yufeng Cui](https://scholar.google.com/citations?hl=en&user=5Ydha2EAAAAJ)<sup>1*</sup>, [Xiaosong Zhang](https://zhangxiaosong18.github.io)<sup>1*</sup>, [Fan Zhang](https://scholar.google.com/citations?user=VsJ39HMAAAAJ)<sup>1*</sup>, [Qiying Yu](https://yqy2001.github.io)<sup>2,1*</sup>, [Zhengxiong Luo](https://greatlog.github.io)<sup>1</sup>, [Yueze Wang]()<sup>1</sup>, [Yongming Rao](https://raoyongming.github.io)<sup>1</sup>,<br>[Jingjing Liu](https://air.tsinghua.edu.cn/en/info/1046/1194.htm)<sup>2</sup>, [Tiejun Huang](https://scholar.google.com/citations?user=knvEK4AAAAAJ&hl=en)<sup>1,3</sup>, [Xinlong Wang](https://www.xloong.wang/)<sup>1†</sup>
11
-
12
- <sup>1</sup> [BAAI](https://www.baai.ac.cn/english.html), <sup>2</sup> [THU](https://air.tsinghua.edu.cn), <sup>3</sup> [PKU](https://english.pku.edu.cn/) <br><sup>*</sup> equal contribution <sup>†</sup> project lead
13
 
14
- | [Paper](https://arxiv.org/abs/2312.13286) | [🤗HF Demo](https://huggingface.co/spaces/BAAI/Emu2) | [Demo](https://emu.ssi.plus) | [Project Page](https://baaivision.github.io/emu2/) | [Github](https://github.com/baaivision/Emu)
15
-
16
- </div>
17
-
18
- The human ability to easily solve multimodal tasks in context (i.e., with only a few demonstrations or simple instructions), is what current multimodal systems have largely struggled to imitate.
19
- In this work, we demonstrate that the task-agnostic in-context learning capabilities of large multimodal models can be significantly enhanced by effective scaling-up.
20
- We introduce **Emu2**, a generative multimodal model with 37 billion parameters, trained on large-scale multimodal sequences with a unified autoregressive objective.
21
- **Emu2** exhibits strong multimodal in-context learning abilities, even emerging to solve tasks that require on-the-fly reasoning, such as visual prompting and object-grounded generation.
22
- The model sets a new record on multiple multimodal understanding tasks in few-shot settings.
23
- When instruction-tuned to follow specific instructions, **Emu2** further achieves new state-of-the-art on challenging tasks such as question answering benchmarks for large multimodal models and open-ended subject-driven generation.
24
- These achievements demonstrate that **Emu2** can serve as a base model and general-purpose interface for a wide range of multimodal tasks.
25
- Code and models are publicly available to facilitate future research.
26
-
27
- # Emu2-Gen
28
 
29
  ## Model Weights
30
 
 
3
  - en
4
  ---
5
 
6
+ # Emu2-Chat
 
 
7
 
8
+ [Paper](https://arxiv.org/abs/2312.13286) | [🤗HF Demo](https://huggingface.co/spaces/BAAI/Emu2) | [Demo](https://emu.ssi.plus) | [Project Page](https://baaivision.github.io/emu2/) | [Github](https://github.com/baaivision/Emu)
 
 
9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
 
11
  ## Model Weights
12