AdaptLLM commited on
Commit
da73565
·
verified ·
1 Parent(s): 5989901

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -3
README.md CHANGED
@@ -6,8 +6,6 @@ language:
6
 
7
  This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
8
 
9
- ## Contributions
10
-
11
  ### 1. Data Synthesis
12
  - We create a **generate-then-filter pipeline** using open-source models to make diverse visual tasks from domain-specific image-caption pairs.
13
  - This data works better than data made by hand or closed-source models (e.g., GPT-4V).
@@ -16,7 +14,7 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
16
  - Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-step training** to handle more tasks for specific domains.
17
 
18
  ### 3. Task Evaluation
19
- - We test our method in important fields like biomedicine, food, and remote sensing.
20
  - We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
21
 
22
 
 
6
 
7
  This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
8
 
 
 
9
  ### 1. Data Synthesis
10
  - We create a **generate-then-filter pipeline** using open-source models to make diverse visual tasks from domain-specific image-caption pairs.
11
  - This data works better than data made by hand or closed-source models (e.g., GPT-4V).
 
14
  - Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-step training** to handle more tasks for specific domains.
15
 
16
  ### 3. Task Evaluation
17
+ - We test our method in important fields like **biomedicine, food, and remote sensing**.
18
  - We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
19
 
20