--- license: apache-2.0 pipeline_tag: image-text-to-text ---

# Spatial-LLaVA-7B Model Card **[Github Repo](https://github.com/xi-jiajun/Spatial-LLaVA)** **[🤗 Huggingface Space Demo](https://huggingface.co/spaces/rogerxi/Spatial-LLaVA)** ## 🤖 Model details **Model type:** This finetuned LLaVA model is trained from [liuhaotian/llava-pretrain-vicuna-7b-v1.3](https://huggingface.co/liuhaotian/llava-pretrain-vicuna-7b-v1.3) for improving spatial relation reasoning of large multi-modal model. LLaVA is an open-source chatbot trained by fine-tuning LLaMA/Vicuna on GPT-generated multimodal instruction-following data. It is an auto-regressive language model, based on the transformer architecture. ## 🎯 Intended use **Primary intended uses:** The primary use of LLaVA is research on large multimodal models and chatbots. **Primary intended users:** The primary intended users of the model are researchers and hobbyists in computer vision, natural language processing, machine learning, and artificial intelligence. ## 📚 Training dataset Instruction following training: [rogerxi/LLaVA-Spatial-Instruct-850K](https://huggingface.co/datasets/rogerxi/LLaVA-Spatial-Instruct-850K) ## 📊 Evaluation A collection of 10 benchmarks: | Model | VQAv2 | GQA | VizWiz | SQA | TextVQA | POPE | MME | MM-Bench | MM-Bench-cn | MM-Vet | |:-----------------------:|:--------:|:--------:|:--------:|:--------:|:--------:|:--------:|:----------:|:--------:|:-----------:|:--------:| | LLaVA-1.5-7b | 78.5 | 62.0 | **50.0** | 66.8 | 58.2 | 85.9 | **1510.7** | 64.3 | 58.3 | 31.1 | | Spatial-LLaVA-7b | **79.7** | **62.7** | 48.7 | **68.7** | **58.5** | **87.2** | 1472.7 | **67.8** | **60.7** | **31.6** | [Spatial-Relation-Eval](https://huggingface.co/datasets/rogerxi/Spatial-Relation-Eval) (built based on [SpatialRGPT-Bench](https://huggingface.co/datasets/a8cheng/SpatialRGPT-Bench)): ### Qualitative Spatial Relations | Model | Below/Above | Left/Right | Big/Small | Tall/Short | Wide/Thin | Behind/Front | Avg | |:-----------------------:|:------------:|:-----------:|:----------:|:-----------:|:----------:|:-------------:|:-------------: | | LLaVA-1.5-7b | 53.91 | 53.49 | 45.36 | 40.00 | **50.00** | 51.04 | 48.97 | | LLaVA-1.5-13b | 54.28 | 52.32 | 45.36 | 48.57 | 49.02 | 47.92 | 49.67 | | Spatial-LLaVA-7b | **56.32** | **66.28** | **60.82** | **48.57** | 49.02 | **52.08** | **55.12** | ### Quantitative Spatial Relations | Model | Direct Dist (m / ratio) | Horizontal Dist (m / ratio) | Vertical Dist (m / ratio) | Width (m / ratio) | Height (m / ratio) | Direction (° / ratio) | |:-----------------------:|:------------------------:|:----------------------------:|:--------------------------:|:--------------------------:|:--------------------------:|:--------------------------:| | LLaVA-1.5-7b | 12.90 / 1.06 | 10.68 / 2.03 | 20.79 / 0.94 | **24.19 / 0.50** | 14.29 / 5.27 | 10.23 / 58.33 | | LLaVA-1.5-13b | 13.71 / 0.93 | 10.68 / 3.56 | 16.83 / 0.85 | 15.32 / 0.57 | 17.67 / 5.8 | 14.77 / 54.29 | | Spatial-LLaVA-7b | **24.19 / 0.57** | **14.56 / 0.62** | **41.58 / 0.42** | 22.58 / 1.12 | **18.25 / 2.92** | **20.45 / 56.47** | ## 🙏 Acknowledgements We thank Liu Haotian et al. for the LLaVA pretrained script, weights and LLaVA-v1.5 mixture dataset; the teams behind CLEVR, TextCaps, VisualMRC and VQAv2 (via “HuggingFaceM4/the_cauldron”); remyxai for OpenSpaces; Anjie Cheng et al. for Spatial-Bench and data pipeline; Google for OpenImages; and Hugging Face for their datasets infrastructure.