Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -13,9 +13,12 @@ tags:
|
|
| 13 |
# OpenAI-Clip: Optimized for Mobile Deployment
|
| 14 |
## Multi-modal foundational model for vision and language tasks like image/text similarity and for zero-shot image classification
|
| 15 |
|
|
|
|
| 16 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
| 17 |
|
| 18 |
-
This model is an implementation of
|
|
|
|
|
|
|
| 19 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|
| 20 |
More details on model performance across various devices, can be found
|
| 21 |
[here](https://aihub.qualcomm.com/models/openai_clip).
|
|
|
|
| 13 |
# OpenAI-Clip: Optimized for Mobile Deployment
|
| 14 |
## Multi-modal foundational model for vision and language tasks like image/text similarity and for zero-shot image classification
|
| 15 |
|
| 16 |
+
|
| 17 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
| 18 |
|
| 19 |
+
This model is an implementation of Posenet-Mobilenet found [here](https://github.com/openai/CLIP/).
|
| 20 |
+
|
| 21 |
+
|
| 22 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|
| 23 |
More details on model performance across various devices, can be found
|
| 24 |
[here](https://aihub.qualcomm.com/models/openai_clip).
|