Update README.md
Browse files
README.md
CHANGED
|
@@ -10,46 +10,43 @@ tags:
|
|
| 10 |
- mllm
|
| 11 |
---
|
| 12 |
|
| 13 |
-
# CrystalChat-7B-MLLM: a fully-reproducible vision
|
| 14 |
|
| 15 |
-
|
| 16 |
|
| 17 |
-
|
| 18 |
-
|-----------------------------------|---------|--------|-------|--------|---------|
|
| 19 |
-
| CrystalCoder-7B | 1359.83 | 238.92 | 86.18 | 64.15 | 50.39 |
|
| 20 |
-
| CrystalChat-7B | 1456.53 | **308.21** | 86.96 | 67.77 | **57.84** |
|
| 21 |
-
| Vicuna-7B | **1481.12** | 302.85 | **87.17** | **67.97** | 56.49 |
|
| 22 |
-
|
| 23 |
-
*Table: Comparison of different LLM backbones on visual language understanding benchmarks. All models are instruction-tuned on the general domain data (i.e. LLaVA)*
|
| 24 |
|
| 25 |
-
|
| 26 |
-
## About Crystal:
|
| 27 |
* 7 billion parameter LLM
|
| 28 |
-
* CLIP ViT-L/14
|
| 29 |
-
* Tokens: ????
|
| 30 |
* Languages: English
|
| 31 |
-
* Models Released:
|
| 32 |
* Trained in 2 stages
|
| 33 |
* License: ?
|
| 34 |
|
| 35 |
-
Crystal-based models were developed as a collaboration between [MBZUAI](https://mbzuai.ac.ae/institute-of-foundation-models/), [Petuum](https://www.petuum.com/), and [LLM360](https://www.llm360.ai/)????.
|
|
|
|
| 36 |
|
| 37 |
## Evaluation
|
| 38 |
|
| 39 |
General Evaluation Metrics for MLLMs. MME serves as an extensive evaluative benchmark,
|
| 40 |
aiming to assess perceptual and cognitive capability of MLLMs within 14 sub-tasks. Additionally, we also evaluate the performance of our models on text-oriented visual question answering tasks employing a diverse set of benchmark datasets including ScienceQA and TextVQA. Furthermore, we assess our models’ ability toward anti-hallucination through POPE.
|
| 41 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
-
|
| 44 |
-
|
| 45 |
|
| 46 |
|
| 47 |
-
##
|
| 48 |
|
| 49 |
### Pretrain Data
|
| 50 |
LLaVA Visual Instruct Pretrain LCS-558K is a filtered subset of the LAION, CC, and SBU datasets, featuring a more balanced distribution of concept coverage. The file includes multimodal synthesized conversations generated from image-caption pairs by incorporating randomly selected instructions such as "Describe this image." It is used for pretraining in LLaVA, with the raw CC-3M caption serving as the default answer.
|
| 51 |
|
| 52 |
-
### Finetune
|
| 53 |
|
| 54 |
The dataset chosen was created by LLaVA with academic-task-oriented VQA data mixture and data from ShareGPT. LLaVA Visual Instruct 150K is a dataset of GPT-generated multimodal instruction-following data. It is designed for visual instruction tuning and aims to develop large multimodal models with capabilities akin to GPT-4 in both vision and language.
|
| 55 |
|
|
@@ -69,12 +66,10 @@ The dataset chosen was created by LLaVA with academic-task-oriented VQA data mix
|
|
| 69 |
| VG [25] | 86K | Provide the bounding box coordinate of the region this sentence describes. |
|
| 70 |
| **Total** | **665K** | |
|
| 71 |
|
| 72 |
-
|
| 73 |
|
| 74 |
|
| 75 |
-
|
| 76 |
-
# LLM360 Research Suite
|
| 77 |
-
|
| 78 |
## Stage 2 - Finetuning
|
| 79 |
| Checkpoints | |
|
| 80 |
| ----------- | ----------- |
|
|
@@ -87,9 +82,16 @@ The dataset chosen was created by LLaVA with academic-task-oriented VQA data mix
|
|
| 87 |
| [CrystalChat](https://huggingface.co/qazimbhat1/Crystal-based-MLLM-7B/tree/Crystal-based-MLLM-7B-pretrain) |
|
| 88 |
| [CrystalCoder](https://huggingface.co/qazimbhat1/Crystal-based-MLLM-7B/tree/Crystal-coder-7B-pretrain) |
|
| 89 |
|
|
|
|
| 90 |
[to find all branches: git branch -a]
|
| 91 |
|
| 92 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 93 |
```python
|
| 94 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 95 |
|
|
@@ -118,6 +120,7 @@ We believe in a future where artificial general intelligence (AGI) is created by
|
|
| 118 |
|
| 119 |
[Visit us](https://www.llm360.ai/)
|
| 120 |
|
|
|
|
| 121 |
## Citation
|
| 122 |
|
| 123 |
**BibTeX:**
|
|
|
|
| 10 |
- mllm
|
| 11 |
---
|
| 12 |
|
| 13 |
+
# CrystalChat-7B-MLLM: a fully-reproducible vision language model based on CrystalChat-7B
|
| 14 |
|
| 15 |
+
## Model Description
|
| 16 |
|
| 17 |
+
CrystalChat-7B based multi-modal large language model (MLLM) mimics the training recipe used for Vicuna-7B based [LLaVa-v1.5](https://huggingface.co/docs/transformers/main/model_doc/llava). CrystalChat-7B-MLLM models are entirely transparent, having open-sourced all materials, including code, data, model checkpoint, intermediate results, and more at [TODO: Add paper link]().
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
|
| 19 |
+
### About CrystalChat-7B-MLLM:
|
|
|
|
| 20 |
* 7 billion parameter LLM
|
| 21 |
+
* CLIP ViT-L/14-336px vision encoder
|
|
|
|
| 22 |
* Languages: English
|
| 23 |
+
* Models Released: CrystalChat-7B-MLLM
|
| 24 |
* Trained in 2 stages
|
| 25 |
* License: ?
|
| 26 |
|
| 27 |
+
Crystal-based models were developed as a collaboration between [MBZUAI](https://mbzuai.ac.ae/institute-of-foundation-models/), [Petuum](https://www.petuum.com/), and [LLM360](https://www.llm360.ai/) TODO- check????.
|
| 28 |
+
|
| 29 |
|
| 30 |
## Evaluation
|
| 31 |
|
| 32 |
General Evaluation Metrics for MLLMs. MME serves as an extensive evaluative benchmark,
|
| 33 |
aiming to assess perceptual and cognitive capability of MLLMs within 14 sub-tasks. Additionally, we also evaluate the performance of our models on text-oriented visual question answering tasks employing a diverse set of benchmark datasets including ScienceQA and TextVQA. Furthermore, we assess our models’ ability toward anti-hallucination through POPE.
|
| 34 |
|
| 35 |
+
| LLM Backbone | MME-P | MME-C | POPE | SciQA | TextVQA |
|
| 36 |
+
|-----------------------------------|---------|--------|-------|--------|---------|
|
| 37 |
+
| CrystalCoder-7B | 1359.83 | 238.92 | 86.18 | 64.15 | 50.39 |
|
| 38 |
+
| CrystalChat-7B | 1456.53 | **308.21** | 86.96 | 67.77 | **57.84** |
|
| 39 |
+
| Vicuna-7B | **1481.12** | 302.85 | **87.17** | **67.97** | 56.49 |
|
| 40 |
|
| 41 |
+
*Table 1: Comparison of different LLM backbones on visual language understanding benchmarks. All models are instruction-tuned on the general domain data (i.e. LLaVA)*
|
|
|
|
| 42 |
|
| 43 |
|
| 44 |
+
## Data and Training Details
|
| 45 |
|
| 46 |
### Pretrain Data
|
| 47 |
LLaVA Visual Instruct Pretrain LCS-558K is a filtered subset of the LAION, CC, and SBU datasets, featuring a more balanced distribution of concept coverage. The file includes multimodal synthesized conversations generated from image-caption pairs by incorporating randomly selected instructions such as "Describe this image." It is used for pretraining in LLaVA, with the raw CC-3M caption serving as the default answer.
|
| 48 |
|
| 49 |
+
### Finetune Data
|
| 50 |
|
| 51 |
The dataset chosen was created by LLaVA with academic-task-oriented VQA data mixture and data from ShareGPT. LLaVA Visual Instruct 150K is a dataset of GPT-generated multimodal instruction-following data. It is designed for visual instruction tuning and aims to develop large multimodal models with capabilities akin to GPT-4 in both vision and language.
|
| 52 |
|
|
|
|
| 66 |
| VG [25] | 86K | Provide the bounding box coordinate of the region this sentence describes. |
|
| 67 |
| **Total** | **665K** | |
|
| 68 |
|
| 69 |
+
*Table 2. Instruction-following Data Mixture of LLaVA-1.5.*
|
| 70 |
|
| 71 |
|
| 72 |
+
TODO: Check if we need to publish these 2
|
|
|
|
|
|
|
| 73 |
## Stage 2 - Finetuning
|
| 74 |
| Checkpoints | |
|
| 75 |
| ----------- | ----------- |
|
|
|
|
| 82 |
| [CrystalChat](https://huggingface.co/qazimbhat1/Crystal-based-MLLM-7B/tree/Crystal-based-MLLM-7B-pretrain) |
|
| 83 |
| [CrystalCoder](https://huggingface.co/qazimbhat1/Crystal-based-MLLM-7B/tree/Crystal-coder-7B-pretrain) |
|
| 84 |
|
| 85 |
+
|
| 86 |
[to find all branches: git branch -a]
|
| 87 |
|
| 88 |
+
## Examples
|
| 89 |
+
|
| 90 |
+
TODO: Add image as sample example
|
| 91 |
+
<center><img src="k2_table_of_tables.png" alt="k2 big eval table"/></center>
|
| 92 |
+
|
| 93 |
+
|
| 94 |
+
## Loading Crystal
|
| 95 |
```python
|
| 96 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 97 |
|
|
|
|
| 120 |
|
| 121 |
[Visit us](https://www.llm360.ai/)
|
| 122 |
|
| 123 |
+
|
| 124 |
## Citation
|
| 125 |
|
| 126 |
**BibTeX:**
|