Image Feature Extraction
Transformers
OpenCLIP
How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("image-feature-extraction", model="UCSC-VLAA/openvision-vit-base-patch8-384")
# Load model directly
from transformers import AutoModel
model = AutoModel.from_pretrained("UCSC-VLAA/openvision-vit-base-patch8-384", dtype="auto")
Quick Links

This repository contains the OpenVision model, a fully-open and cost-effective family of advanced vision encoders for multimodal learning, as described in the paper OpenVision: A Fully-Open, Cost-Effective Family of Advanced Vision Encoders for Multimodal Learning.

Project Page: https://ucsc-vlaa.github.io/OpenVision/

Code: https://github.com/UCSC-VLAA/OpenVision

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Collection including UCSC-VLAA/openvision-vit-base-patch8-384

Paper for UCSC-VLAA/openvision-vit-base-patch8-384