Spaces:
Runtime error
Runtime error
added prefix for IC
Browse files
app.py
CHANGED
|
@@ -243,6 +243,8 @@ def inference(img, text, is_translate):
|
|
| 243 |
text = clip.tokenize([text]).to(device)
|
| 244 |
with torch.no_grad():
|
| 245 |
prefix = clip_model.encode_text(text).to(device, dtype=torch.float32)
|
|
|
|
|
|
|
| 246 |
|
| 247 |
else:
|
| 248 |
if img is None:
|
|
@@ -253,13 +255,9 @@ def inference(img, text, is_translate):
|
|
| 253 |
|
| 254 |
with torch.no_grad():
|
| 255 |
prefix = clip_model.encode_image(image).to(device, dtype=torch.float32)
|
|
|
|
|
|
|
| 256 |
|
| 257 |
-
prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
|
| 258 |
-
|
| 259 |
-
if use_beam_search:
|
| 260 |
-
generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed)[0]
|
| 261 |
-
else:
|
| 262 |
-
generated_text_prefix = generate2(model, tokenizer, embed=prefix_embed)
|
| 263 |
return generated_text_prefix
|
| 264 |
|
| 265 |
title = "CLIP Dual encoder"
|
|
|
|
| 243 |
text = clip.tokenize([text]).to(device)
|
| 244 |
with torch.no_grad():
|
| 245 |
prefix = clip_model.encode_text(text).to(device, dtype=torch.float32)
|
| 246 |
+
prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
|
| 247 |
+
generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed)[0]
|
| 248 |
|
| 249 |
else:
|
| 250 |
if img is None:
|
|
|
|
| 255 |
|
| 256 |
with torch.no_grad():
|
| 257 |
prefix = clip_model.encode_image(image).to(device, dtype=torch.float32)
|
| 258 |
+
prefix_embed = model.clip_project(prefix).reshape(1, prefix_length, -1)
|
| 259 |
+
generated_text_prefix = generate_beam(model, tokenizer, embed=prefix_embed, prompt="Một bức ảnh về")[0]
|
| 260 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 261 |
return generated_text_prefix
|
| 262 |
|
| 263 |
title = "CLIP Dual encoder"
|