Instructions to use Salesforce/blip-image-captioning-base with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use Salesforce/blip-image-captioning-base with Transformers:
# Use a pipeline as a high-level helper # Warning: Pipeline type "image-to-text" is no longer supported in transformers v5. # You must load the model directly (see below) or downgrade to v4.x with: # 'pip install "transformers<5.0.0' from transformers import pipeline pipe = pipeline("image-to-text", model="Salesforce/blip-image-captioning-base")# Load model directly from transformers import AutoProcessor, AutoModelForImageTextToText processor = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-base") model = AutoModelForImageTextToText.from_pretrained("Salesforce/blip-image-captioning-base") - Notebooks
- Google Colab
- Kaggle
Update README.md
#5
by patrickvonplaten - opened
README.md
CHANGED
|
@@ -37,8 +37,8 @@ import requests
|
|
| 37 |
from PIL import Image
|
| 38 |
from transformers import BlipProcessor, BlipForConditionalGeneration
|
| 39 |
|
| 40 |
-
processor = BlipProcessor.from_pretrained("
|
| 41 |
-
model = BlipForConditionalGeneration.from_pretrained("
|
| 42 |
|
| 43 |
img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg'
|
| 44 |
raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')
|
|
|
|
| 37 |
from PIL import Image
|
| 38 |
from transformers import BlipProcessor, BlipForConditionalGeneration
|
| 39 |
|
| 40 |
+
processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-base")
|
| 41 |
+
model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-base")
|
| 42 |
|
| 43 |
img_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg'
|
| 44 |
raw_image = Image.open(requests.get(img_url, stream=True).raw).convert('RGB')
|