Add links to Github repository, project page and dataset
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,8 +1,9 @@
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
-
pipeline_tag: image-text-to-text
|
| 4 |
license: apache-2.0
|
|
|
|
| 5 |
---
|
|
|
|
| 6 |
# Model Card: Reflective LLaVA (ReflectiVA)
|
| 7 |
|
| 8 |
Multimodal LLMs (MLLMs) are the natural extension of large language models to handle multimodal inputs, combining text and image data.
|
|
@@ -20,7 +21,7 @@ superior performance compared to existing methods.
|
|
| 20 |
|
| 21 |
In this model space, you will find the Overall Model (stage two) weights of ```ReflectiVA```.
|
| 22 |
|
| 23 |
-
For more information, visit our [ReflectiVA repository](https://github.com/aimagelab/ReflectiVA).
|
| 24 |
|
| 25 |
## Citation
|
| 26 |
If you make use of our work, please cite our repo:
|
|
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
|
|
|
| 3 |
license: apache-2.0
|
| 4 |
+
pipeline_tag: image-text-to-text
|
| 5 |
---
|
| 6 |
+
|
| 7 |
# Model Card: Reflective LLaVA (ReflectiVA)
|
| 8 |
|
| 9 |
Multimodal LLMs (MLLMs) are the natural extension of large language models to handle multimodal inputs, combining text and image data.
|
|
|
|
| 21 |
|
| 22 |
In this model space, you will find the Overall Model (stage two) weights of ```ReflectiVA```.
|
| 23 |
|
| 24 |
+
For more information, visit our [ReflectiVA repository](https://github.com/aimagelab/ReflectiVA), our [project page](https://aimagelab.github.io/ReflectiVA/) and the [dataset](https://huggingface.co/datasets/aimagelab/ReflectiVA-Data).
|
| 25 |
|
| 26 |
## Citation
|
| 27 |
If you make use of our work, please cite our repo:
|