Update README.md
Browse files
README.md
CHANGED
|
@@ -12,6 +12,8 @@ Disclaimer: The team releasing SigLIP did not write a model card for this model
|
|
| 12 |
|
| 13 |
SigLIP is [CLIP](https://huggingface.co/docs/transformers/model_doc/clip), a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a global view of the pairwise similarities for normalization. This allows further scaling up the batch size, while also performing better at smaller batch sizes.
|
| 14 |
|
|
|
|
|
|
|
| 15 |
## Intended uses & limitations
|
| 16 |
|
| 17 |
You can use the raw model for tasks like zero-shot image classification and image-text retrieval. See the [model hub](https://huggingface.co/models?search=google/siglip) to look for
|
|
|
|
| 12 |
|
| 13 |
SigLIP is [CLIP](https://huggingface.co/docs/transformers/model_doc/clip), a multimodal model, with a better loss function. The sigmoid loss operates solely on image-text pairs and does not require a global view of the pairwise similarities for normalization. This allows further scaling up the batch size, while also performing better at smaller batch sizes.
|
| 14 |
|
| 15 |
+
A TLDR of SigLIP by one of the authors can be found [here](https://twitter.com/giffmana/status/1692641733459267713).
|
| 16 |
+
|
| 17 |
## Intended uses & limitations
|
| 18 |
|
| 19 |
You can use the raw model for tasks like zero-shot image classification and image-text retrieval. See the [model hub](https://huggingface.co/models?search=google/siglip) to look for
|