Instructions to use jinaai/jina-clip-implementation with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use jinaai/jina-clip-implementation with Transformers:
# Load model directly from transformers import AutoModel model = AutoModel.from_pretrained("jinaai/jina-clip-implementation", dtype="auto") - Notebooks
- Google Colab
- Kaggle
Jina CLIP
Core implementation of Jina CLIP. The model uses:
- the EVA 02 architecture for the vision tower
- the Jina XLM RoBERTa with Flash Attention model as a text tower
Models that use this implementation
Requirements
To use the Jina CLIP source code, the following packages are required:
torchtimmtransformerseinopsxformersto use x-attentionflash-attnto use flash attentionapexto use fused layer normalization
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support