Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -2,7 +2,6 @@ import gradio as gr
|
|
| 2 |
from transformers import Blip2Processor, Blip2ForConditionalGeneration
|
| 3 |
import torch
|
| 4 |
import PIL.Image
|
| 5 |
-
from huggingface_hub import HfApi
|
| 6 |
|
| 7 |
BLIP_MODEL_ID = "Salesforce/blip2-opt-2.7b"
|
| 8 |
|
|
@@ -15,9 +14,6 @@ if device != 'cuda':
|
|
| 15 |
blip_processor = Blip2Processor.from_pretrained(BLIP_MODEL_ID)
|
| 16 |
blip_model = Blip2ForConditionalGeneration.from_pretrained(BLIP_MODEL_ID, torch_dtype=torch.float16, device_map="auto").to(device)
|
| 17 |
|
| 18 |
-
print("Loading API")
|
| 19 |
-
api = HfApi()
|
| 20 |
-
|
| 21 |
|
| 22 |
def caption_images(images: list[PIL.Image.Image], prompt: str) -> dict[str, str]:
|
| 23 |
image_files = [PIL.Image.open(image.name).convert('RGB') for image in images]
|
|
|
|
| 2 |
from transformers import Blip2Processor, Blip2ForConditionalGeneration
|
| 3 |
import torch
|
| 4 |
import PIL.Image
|
|
|
|
| 5 |
|
| 6 |
BLIP_MODEL_ID = "Salesforce/blip2-opt-2.7b"
|
| 7 |
|
|
|
|
| 14 |
blip_processor = Blip2Processor.from_pretrained(BLIP_MODEL_ID)
|
| 15 |
blip_model = Blip2ForConditionalGeneration.from_pretrained(BLIP_MODEL_ID, torch_dtype=torch.float16, device_map="auto").to(device)
|
| 16 |
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
def caption_images(images: list[PIL.Image.Image], prompt: str) -> dict[str, str]:
|
| 19 |
image_files = [PIL.Image.open(image.name).convert('RGB') for image in images]
|