Update README.md
Browse files
README.md
CHANGED
|
@@ -1,3 +1,15 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
tags:
|
| 4 |
+
- vision
|
| 5 |
+
- vision-language-model
|
| 6 |
+
- contrastive learning
|
| 7 |
+
- self-supervised learning
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
**COSMOS Model**
|
| 11 |
+
|
| 12 |
+
Authors: [Sanghwan Kim](https://kim-sanghwan.github.io/), [Rui Xiao](https://www.eml-munich.de/people/rui-xiao), [Mariana-Iuliana Georgescu](https://lilygeorgescu.github.io/), [Stephan Alaniz](https://www.eml-munich.de/people/stephan-alaniz), [Zeynep Akata](https://www.eml-munich.de/people/zeynep-akata)
|
| 13 |
+
|
| 14 |
+
COSMOS is introduced in the paper [COSMOS: Cross-Modality Self-Distillation for Vision Language Pre-training](https://arxiv.org/abs/2412.01814). COSMOS is trained in self-supervised learning framework with multi-modal augmentation and cross-attention module. It outperforms CLIP-based models trained on larger datasets in visual perception and contextual understanding tasks. COSMOS achieves strong performance in downstream tasks including zero-shot image-text retrieval, classification, and semantic segmentation segmentation.
|
| 15 |
+
|