Update README.md
Browse files
README.md
CHANGED
|
@@ -6,8 +6,6 @@ language:
|
|
| 6 |
|
| 7 |
This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
|
| 8 |
|
| 9 |
-
## Contributions
|
| 10 |
-
|
| 11 |
### 1. Data Synthesis
|
| 12 |
- We create a **generate-then-filter pipeline** using open-source models to make diverse visual tasks from domain-specific image-caption pairs.
|
| 13 |
- This data works better than data made by hand or closed-source models (e.g., GPT-4V).
|
|
@@ -16,7 +14,7 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
|
|
| 16 |
- Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-step training** to handle more tasks for specific domains.
|
| 17 |
|
| 18 |
### 3. Task Evaluation
|
| 19 |
-
- We test our method in important fields like biomedicine, food, and remote sensing.
|
| 20 |
- We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
|
| 21 |
|
| 22 |
|
|
|
|
| 6 |
|
| 7 |
This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
|
| 8 |
|
|
|
|
|
|
|
| 9 |
### 1. Data Synthesis
|
| 10 |
- We create a **generate-then-filter pipeline** using open-source models to make diverse visual tasks from domain-specific image-caption pairs.
|
| 11 |
- This data works better than data made by hand or closed-source models (e.g., GPT-4V).
|
|
|
|
| 14 |
- Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-step training** to handle more tasks for specific domains.
|
| 15 |
|
| 16 |
### 3. Task Evaluation
|
| 17 |
+
- We test our method in important fields like **biomedicine, food, and remote sensing**.
|
| 18 |
- We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
|
| 19 |
|
| 20 |
|