Update app.py
Browse files
app.py
CHANGED
|
@@ -290,7 +290,6 @@ def _get_scenes_extraction(
|
|
| 290 |
scene_info: List[Dict] = []
|
| 291 |
|
| 292 |
for i, (start_time, end_time) in enumerate(scene_list):
|
| 293 |
-
print("a")
|
| 294 |
frame_number = int(start_time.get_frames()) + offset_frames
|
| 295 |
cap.set(cv2.CAP_PROP_POS_FRAMES, frame_number)
|
| 296 |
ret, frame = cap.read()
|
|
@@ -343,7 +342,8 @@ def _get_image_list_description(
|
|
| 343 |
model = LlavaOnevisionForConditionalGeneration.from_pretrained(
|
| 344 |
path_model,
|
| 345 |
torch_dtype=torch.float16,
|
| 346 |
-
low_cpu_mem_usage=
|
|
|
|
| 347 |
).to("cuda")
|
| 348 |
|
| 349 |
# System prompt for image description
|
|
|
|
| 290 |
scene_info: List[Dict] = []
|
| 291 |
|
| 292 |
for i, (start_time, end_time) in enumerate(scene_list):
|
|
|
|
| 293 |
frame_number = int(start_time.get_frames()) + offset_frames
|
| 294 |
cap.set(cv2.CAP_PROP_POS_FRAMES, frame_number)
|
| 295 |
ret, frame = cap.read()
|
|
|
|
| 342 |
model = LlavaOnevisionForConditionalGeneration.from_pretrained(
|
| 343 |
path_model,
|
| 344 |
torch_dtype=torch.float16,
|
| 345 |
+
low_cpu_mem_usage=True,
|
| 346 |
+
device_map="auto"
|
| 347 |
).to("cuda")
|
| 348 |
|
| 349 |
# System prompt for image description
|