Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -73,7 +73,7 @@ feature_extractor = ViTImageProcessor.from_pretrained(encoder_name_or_path)
|
|
| 73 |
|
| 74 |
|
| 75 |
|
| 76 |
-
|
| 77 |
|
| 78 |
# cap_model.to(device)
|
| 79 |
|
|
@@ -101,7 +101,7 @@ def predict_event(image):
|
|
| 101 |
|
| 102 |
# img = Image.open(image).convert("RGB")
|
| 103 |
|
| 104 |
-
generated_caption = tokenizer.decode(model.generate(feature_extractor(image, return_tensors="pt").pixel_values.to(
|
| 105 |
|
| 106 |
# caption_vitgpt = generate_caption(model, image)
|
| 107 |
#caption_vitgpt = generate_caption(feature_extractor, cap_model, image, tokenizer)
|
|
|
|
| 73 |
|
| 74 |
|
| 75 |
|
| 76 |
+
device = "cuda" if torch.cuda.is_available() else "cpu"
|
| 77 |
|
| 78 |
# cap_model.to(device)
|
| 79 |
|
|
|
|
| 101 |
|
| 102 |
# img = Image.open(image).convert("RGB")
|
| 103 |
|
| 104 |
+
generated_caption = tokenizer.decode(model.generate(feature_extractor(image, return_tensors="pt").pixel_values.to(device))[0])
|
| 105 |
|
| 106 |
# caption_vitgpt = generate_caption(model, image)
|
| 107 |
#caption_vitgpt = generate_caption(feature_extractor, cap_model, image, tokenizer)
|