File size: 3,974 Bytes
37fd3cc
 
 
 
 
 
 
 
 
 
dcc0afe
 
4871778
 
37fd3cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
ea864cf
37fd3cc
 
 
e7d3ba3
37fd3cc
 
 
 
 
ef4da75
37fd3cc
 
 
 
 
2fd18e2
37fd3cc
ef4da75
37fd3cc
ea864cf
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: apache-2.0
pipeline_tag: image-text-to-text
---

<br>
<br>

# Spatial-LLaVA-7B Model Card

**[Github Repo](https://github.com/xi-jiajun/Spatial-LLaVA)**

**[🤗 Huggingface Space Demo](https://huggingface.co/spaces/rogerxi/Spatial-LLaVA)**

## 🤖 Model details 

**Model type:**

This finetuned LLaVA model is trained from [liuhaotian/llava-pretrain-vicuna-7b-v1.3](https://huggingface.co/liuhaotian/llava-pretrain-vicuna-7b-v1.3) for improving spatial relation reasoning of large multi-modal model. 

LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data.
It is an auto-regressive language model, based on the transformer architecture.

## 🎯 Intended use 
**Primary intended uses:**
The primary use of LLaVA is research on large multimodal models and chatbots.

**Primary intended users:**
The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence.

## 📚 Training dataset 
Instruction following training: [rogerxi/LLaVA-Spatial-Instruct-850K](https://huggingface.co/datasets/rogerxi/LLaVA-Spatial-Instruct-850K)

## 📊 Evaluation 
A collection of 10 benchmarks:
| Model                  |   VQAv2  |    GQA   |  VizWiz  |    SQA   |  TextVQA |   POPE   |     MME    | MM-Bench | MM-Bench-cn |  MM-Vet  |
|:-----------------------:|:--------:|:--------:|:--------:|:--------:|:--------:|:--------:|:----------:|:--------:|:-----------:|:--------:|
|   LLaVA-1.5-7b   |   78.5   |   62.0   | **50.0** |   66.8   |   58.2   |   85.9   | **1510.7** |   64.3   |     58.3    |   31.1   |
| Spatial-LLaVA-7b | **79.7** | **62.7** |   48.7   | **68.7** | **58.5** | **87.2** |   1472.7   | **67.8** |   **60.7**  | **31.6** |

[Spatial-Relation-Eval](https://huggingface.co/datasets/rogerxi/Spatial-Relation-Eval) (built based on [SpatialRGPT-Bench](https://huggingface.co/datasets/a8cheng/SpatialRGPT-Bench)):
### Qualitative Spatial Relations

| Model                 | Below/Above | Left/Right | Big/Small | Tall/Short | Wide/Thin | Behind/Front | Avg |
|:-----------------------:|:------------:|:-----------:|:----------:|:-----------:|:----------:|:-------------:|:-------------: |
|   LLaVA-1.5-7b        |      53.91 |     53.49 |    45.36 |     40.00 |    **50.00** |     51.04 |  48.97  |
|   LLaVA-1.5-13b        |      54.28 |     52.32 |    45.36 |     48.57 |    49.02 |     47.92 |  49.67  | 
|   Spatial-LLaVA-7b    |      **56.32** |     **66.28** |    **60.82** |     **48.57** |    49.02 |      **52.08** | **55.12** |

### Quantitative Spatial Relations

| Model                 | Direct Dist (m / ratio) | Horizontal Dist (m / ratio) | Vertical Dist (m / ratio) | Width (m / ratio) | Height (m / ratio) | Direction (° / ratio) |
|:-----------------------:|:------------------------:|:----------------------------:|:--------------------------:|:--------------------------:|:--------------------------:|:--------------------------:|
| LLaVA-1.5-7b          |     12.90 / 1.06         |     10.68 / 2.03             |     20.79 / 0.94          |     **24.19 / 0.50**  |     14.29 / 5.27   |      10.23 / 58.33    |
| LLaVA-1.5-13b          |     13.71 / 0.93         |     10.68 / 3.56             |     16.83 / 0.85          |    15.32 / 0.57  |    17.67 / 5.8   |      14.77 / 54.29    |
| Spatial-LLaVA-7b      |     **24.19 / 0.57**         |     **14.56 / 0.62**             |     **41.58 / 0.42**          |     22.58 / 1.12  |     **18.25 / 2.92**   |      **20.45 / 56.47**    |

## 🙏 Acknowledgements
We thank Liu Haotian et al. for the LLaVA pretrained script, weights and LLaVA-v1.5 mixture dataset; the teams behind CLEVR, TextCaps, VisualMRC and VQAv2 (via “HuggingFaceM4/the_cauldron”); remyxai for OpenSpaces; Anjie Cheng et al. for Spatial-Bench and data pipeline; Google for OpenImages; and Hugging Face for their datasets infrastructure.