| import os | |
| import io | |
| from PIL import Image | |
| from transformers import pipeline | |
| import gradio as gr | |
| get_completion = pipeline("image-to-text",model="Salesforce/blip-image-captioning-base") | |
| def summarize(input): | |
| output = get_completion(input) | |
| return output[0]['generated_text'] | |
| def captioner(image): | |
| result = get_completion(image) | |
| return result[0]['generated_text'] | |
| gr.close_all() | |
| christmas_dog = "dog_animal_greyhound_983023.jpg" | |
| bird = "bird_exotic_bird_green.jpg" | |
| cow = "cow_animal_cow_head.jpg" | |
| demo = gr.Interface(fn=captioner, | |
| inputs=[gr.Image(label="Upload image", type="pil", value=christmas_dog)], | |
| outputs=[gr.Textbox(label="Caption")], | |
| title="Image Captioning with BLIP", | |
| description="Caption any image using the BLIP model", | |
| allow_flagging="never", | |
| examples=[christmas_dog, bird, cow]) | |
| demo.launch(share=True) | |