|
|
|
|
|
""" |
|
|
This file extracts the frames for the frame datasets in SA-CO/Gold and Silver. |
|
|
|
|
|
Call like: |
|
|
> python extract_frames.py <dataset_name> |
|
|
""" |
|
|
|
|
|
import json |
|
|
import os |
|
|
import shutil |
|
|
import sys |
|
|
from multiprocessing import Pool |
|
|
|
|
|
from PIL import Image |
|
|
from tqdm import tqdm |
|
|
from utils import ( |
|
|
annotation_files, |
|
|
config, |
|
|
get_frame_from_video, |
|
|
is_valid_image, |
|
|
update_annotations, |
|
|
) |
|
|
|
|
|
|
|
|
def extract_frame(path_video, global_frame_idx, path_frame, image_size, file_name): |
|
|
frame = get_frame_from_video(path_video, global_frame_idx) |
|
|
os.makedirs(os.path.dirname(path_frame), exist_ok=True) |
|
|
img = Image.fromarray(frame) |
|
|
if frame.shape[:2] != image_size: |
|
|
print(f"Resizing image {file_name} from {frame.shape[:2]} to {image_size}") |
|
|
height, width = image_size |
|
|
img = img.resize((width, height)) |
|
|
img.save(path_frame) |
|
|
|
|
|
|
|
|
def process_image(args): |
|
|
image, dataset_name, config = args |
|
|
original_video, global_frame_idx, file_name, image_size = image |
|
|
extra_subpath = "" |
|
|
if dataset_name == "ego4d": |
|
|
extra_subpath = "v1/clips" |
|
|
elif dataset_name == "yt1b": |
|
|
original_video = f"video_{original_video}.mp4" |
|
|
elif dataset_name == "sav": |
|
|
extra_subpath = "videos_fps_6" |
|
|
path_video = os.path.join( |
|
|
config[f"{dataset_name}_path"], |
|
|
"downloaded_videos", |
|
|
extra_subpath, |
|
|
original_video, |
|
|
) |
|
|
path_frame = os.path.join(config[f"{dataset_name}_path"], "frames", file_name) |
|
|
to_return = file_name |
|
|
try: |
|
|
extract_frame(path_video, global_frame_idx, path_frame, image_size, file_name) |
|
|
if not is_valid_image(path_frame): |
|
|
print(f"Invalid image in {path_frame}") |
|
|
to_return = None |
|
|
except: |
|
|
print(f"Invalid image in {path_frame}") |
|
|
to_return = None |
|
|
return to_return |
|
|
|
|
|
|
|
|
def main(): |
|
|
assert len(sys.argv) > 1, "You have to provide the name of the dataset" |
|
|
dataset_name = sys.argv[1] |
|
|
assert ( |
|
|
dataset_name in annotation_files |
|
|
), f"The dataset can be one of {list(annotation_files.keys())}" |
|
|
all_outputs = [] |
|
|
for file in annotation_files[dataset_name]: |
|
|
with open(os.path.join(config["path_annotations"], file), "r") as f: |
|
|
annotation = json.load(f) |
|
|
images = annotation["images"] |
|
|
images = set( |
|
|
( |
|
|
image["original_video"], |
|
|
image["global_frame_idx"], |
|
|
image["file_name"], |
|
|
tuple(image["image_size"]), |
|
|
) |
|
|
for image in images |
|
|
) |
|
|
args_list = [(image, dataset_name, config) for image in images] |
|
|
with Pool(os.cpu_count()) as pool: |
|
|
outputs = list( |
|
|
tqdm(pool.imap_unordered(process_image, args_list), total=len(images)) |
|
|
) |
|
|
all_outputs.extend(outputs) |
|
|
if any(out is None for out in outputs): |
|
|
update_annotations(dataset_name, all_outputs, key="file_name") |
|
|
if config[f"remove_downloaded_videos_{dataset_name}"]: |
|
|
shutil.rmtree(os.path.join(config[f"{dataset_name}_path"], "downloaded_videos")) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|