Instructions to use carsonpoole/binary-siglip-vision with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use carsonpoole/binary-siglip-vision with Transformers:
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("image-feature-extraction", model="carsonpoole/binary-siglip-vision")# Load model directly from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("carsonpoole/binary-siglip-vision") model = AutoModel.from_pretrained("carsonpoole/binary-siglip-vision") - Notebooks
- Google Colab
- Kaggle
Description This is a fine tuned google/siglip-so400m-patch14-384 for the purpose of quantizing the embeddings to binary. It's only using the first 1024 embeddings, so if you use all 1152 of them your results will be less than desirable.
I updated the model today (April 30th) and evals are much better than before, but I'm continuing training so perf should only get better from here.
Evals Coming soon
- Downloads last month
- 8
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support