update README
Browse files
README.md
CHANGED
|
@@ -3,28 +3,10 @@ language:
|
|
| 3 |
- en
|
| 4 |
---
|
| 5 |
|
| 6 |
-
|
| 7 |
-
<h1>Generative Multimodal Models are In-Context Learners</h1h1>
|
| 8 |
-
<h3><a href="">Generative Multimodal Models are In-Context Learners</a></h3>
|
| 9 |
|
| 10 |
-
[
|
| 11 |
-
|
| 12 |
-
<sup>1</sup> [BAAI](https://www.baai.ac.cn/english.html), <sup>2</sup> [THU](https://air.tsinghua.edu.cn), <sup>3</sup> [PKU](https://english.pku.edu.cn/) <br><sup>*</sup> equal contribution <sup>†</sup> project lead
|
| 13 |
|
| 14 |
-
| [Paper](https://arxiv.org/abs/2312.13286) | [🤗HF Demo](https://huggingface.co/spaces/BAAI/Emu2) | [Demo](https://emu.ssi.plus) | [Project Page](https://baaivision.github.io/emu2/) | [Github](https://github.com/baaivision/Emu)
|
| 15 |
-
|
| 16 |
-
</div>
|
| 17 |
-
|
| 18 |
-
The human ability to easily solve multimodal tasks in context (i.e., with only a few demonstrations or simple instructions), is what current multimodal systems have largely struggled to imitate.
|
| 19 |
-
In this work, we demonstrate that the task-agnostic in-context learning capabilities of large multimodal models can be significantly enhanced by effective scaling-up.
|
| 20 |
-
We introduce **Emu2**, a generative multimodal model with 37 billion parameters, trained on large-scale multimodal sequences with a unified autoregressive objective.
|
| 21 |
-
**Emu2** exhibits strong multimodal in-context learning abilities, even emerging to solve tasks that require on-the-fly reasoning, such as visual prompting and object-grounded generation.
|
| 22 |
-
The model sets a new record on multiple multimodal understanding tasks in few-shot settings.
|
| 23 |
-
When instruction-tuned to follow specific instructions, **Emu2** further achieves new state-of-the-art on challenging tasks such as question answering benchmarks for large multimodal models and open-ended subject-driven generation.
|
| 24 |
-
These achievements demonstrate that **Emu2** can serve as a base model and general-purpose interface for a wide range of multimodal tasks.
|
| 25 |
-
Code and models are publicly available to facilitate future research.
|
| 26 |
-
|
| 27 |
-
# Emu2-Gen
|
| 28 |
|
| 29 |
## Model Weights
|
| 30 |
|
|
|
|
| 3 |
- en
|
| 4 |
---
|
| 5 |
|
| 6 |
+
# Emu2-Chat
|
|
|
|
|
|
|
| 7 |
|
| 8 |
+
[Paper](https://arxiv.org/abs/2312.13286) | [🤗HF Demo](https://huggingface.co/spaces/BAAI/Emu2) | [Demo](https://emu.ssi.plus) | [Project Page](https://baaivision.github.io/emu2/) | [Github](https://github.com/baaivision/Emu)
|
|
|
|
|
|
|
| 9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
|
| 11 |
## Model Weights
|
| 12 |
|