mblackman commited on
Commit
7980939
·
1 Parent(s): 2876bfa

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +0 -4
app.py CHANGED
@@ -2,7 +2,6 @@ import gradio as gr
2
  from transformers import Blip2Processor, Blip2ForConditionalGeneration
3
  import torch
4
  import PIL.Image
5
- from huggingface_hub import HfApi
6
 
7
  BLIP_MODEL_ID = "Salesforce/blip2-opt-2.7b"
8
 
@@ -15,9 +14,6 @@ if device != 'cuda':
15
  blip_processor = Blip2Processor.from_pretrained(BLIP_MODEL_ID)
16
  blip_model = Blip2ForConditionalGeneration.from_pretrained(BLIP_MODEL_ID, torch_dtype=torch.float16, device_map="auto").to(device)
17
 
18
- print("Loading API")
19
- api = HfApi()
20
-
21
 
22
  def caption_images(images: list[PIL.Image.Image], prompt: str) -> dict[str, str]:
23
  image_files = [PIL.Image.open(image.name).convert('RGB') for image in images]
 
2
  from transformers import Blip2Processor, Blip2ForConditionalGeneration
3
  import torch
4
  import PIL.Image
 
5
 
6
  BLIP_MODEL_ID = "Salesforce/blip2-opt-2.7b"
7
 
 
14
  blip_processor = Blip2Processor.from_pretrained(BLIP_MODEL_ID)
15
  blip_model = Blip2ForConditionalGeneration.from_pretrained(BLIP_MODEL_ID, torch_dtype=torch.float16, device_map="auto").to(device)
16
 
 
 
 
17
 
18
  def caption_images(images: list[PIL.Image.Image], prompt: str) -> dict[str, str]:
19
  image_files = [PIL.Image.open(image.name).convert('RGB') for image in images]