Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
|
@@ -16,7 +16,7 @@ tags:
|
|
| 16 |
|
| 17 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
| 18 |
|
| 19 |
-
This model is an implementation of
|
| 20 |
|
| 21 |
|
| 22 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|
|
|
|
| 16 |
|
| 17 |
Contrastive Language-Image Pre-Training (CLIP) uses a ViT like transformer to get visual features and a causal language model to get the text features. Both the text and visual features can then be used for a variety of zero-shot learning tasks.
|
| 18 |
|
| 19 |
+
This model is an implementation of OpenAI-Clip found [here](https://github.com/openai/CLIP/).
|
| 20 |
|
| 21 |
|
| 22 |
This repository provides scripts to run OpenAI-Clip on Qualcomm® devices.
|