Spaces:
Runtime error
Runtime error
| import gradio as gr | |
| from transformers import AutoProcessor, AutoModelForCausalLM | |
| import torch | |
| device = "cuda" if torch.cuda.is_available() else "cpu" | |
| checkpoint1 = "ayoubkirouane/git-base-One-Piece" | |
| processor = AutoProcessor.from_pretrained(checkpoint1) | |
| model1 = AutoModelForCausalLM.from_pretrained(checkpoint1) | |
| def img2cap(image): | |
| input1 = processor(images=image, return_tensors="pt").to(device) | |
| pixel_values1 = input1.pixel_values | |
| generated_id1 = model1.generate(pixel_values=pixel_values1, max_length=50) | |
| generated_caption1 = processor.batch_decode(generated_id1, skip_special_tokens=True)[0] | |
| return generated_caption1 | |
| examples = ["1.png", "2.png", "3.png"] | |
| gr.Interface( | |
| fn=img2cap, | |
| inputs=gr.Image(type="pil", label="Original Image"), | |
| outputs=gr.Textbox(label="The generated caption:"), | |
| title="Image Captioning using git-base-One-Piece Model", | |
| description="git-base-One-Piece is used to generate Image Caption for the uploaded image.", | |
| examples=examples, | |
| theme="huggingface", | |
| ).launch(debug=True) | |