# Copyright (c) Meta Platforms, Inc. and affiliates. All Rights Reserved import argparse import logging import multiprocessing as mp import os from functools import partial import pandas as pd from saco_yt1b_frame_prep_util import YtVideoPrep from tqdm import tqdm logger = logging.getLogger(__name__) def download_and_extract_frames(saco_yt1b_id, args): video_prep = YtVideoPrep( saco_yt1b_id=saco_yt1b_id, data_dir=args.data_dir, cookies_file=args.cookies_file, yt1b_start_end_time_file=args.yt1b_start_end_time_file, ffmpeg_timeout=args.ffmpeg_timeout, sleep_interval=args.sleep_interval, max_sleep_interval=args.max_sleep_interval, ) status = video_prep.download_youtube_video() logger.info(f"[video download][{saco_yt1b_id}] download status {status}") if status not in ["already exists", "success"]: logger.warning( f"Video download failed for {saco_yt1b_id}, skipping frame generation" ) return False status = video_prep.extract_frames_in_6fps_and_width_1080() logger.info(f"[frame extracting][{saco_yt1b_id}] frame extracting status {status}") return True def main(): parser = argparse.ArgumentParser() parser.add_argument( "--data_dir", type=str, required=True, ) parser.add_argument( "--cookies_file", type=str, required=True, ) parser.add_argument( "--yt1b_start_end_time_file", type=str, required=True, ) parser.add_argument( "--yt1b_frame_prep_log_file", type=str, required=True, ) parser.add_argument( "--ffmpeg_timeout", type=str, default=7200, # Use longer timeout in case of large videos processing timeout ) parser.add_argument( "--sleep_interval", type=int, default=10, ) parser.add_argument( "--max_sleep_interval", type=int, default=30, ) parser.add_argument( "--num_workers", type=int, default=4, ) args = parser.parse_args() log_dir = os.path.dirname(args.yt1b_frame_prep_log_file) if log_dir: os.makedirs(log_dir, exist_ok=True) # Set up logging to both file and console # Configure the ROOT logger so all child loggers inherit the configuration logging.basicConfig( level=logging.INFO, format="%(asctime)s [%(processName)s/%(threadName)s] %(name)s - %(levelname)s: %(message)s", handlers=[ logging.FileHandler(args.yt1b_frame_prep_log_file, mode="w"), logging.StreamHandler(), ], force=True, # Override any existing configuration ) YT_DLP_WARNING_STR = """ ========== NOTICE!! This script uses yt-dlp to download youtube videos. See the youtube account banning risk in https://github.com/yt-dlp/yt-dlp/wiki/Extractors#exporting-youtube-cookies ========== """ logger.info(YT_DLP_WARNING_STR) args = parser.parse_args() with open(args.yt1b_start_end_time_file, "r") as f: yt1b_start_end_time_df = pd.read_json(f) saco_yt1b_ids = yt1b_start_end_time_df.saco_yt1b_id.unique() num_workers = args.num_workers logger.info( f"Starting with {num_workers} parallel worker(s) (sleep_interval={args.sleep_interval}-{args.max_sleep_interval}s)" ) with mp.Pool(num_workers) as p: download_func = partial(download_and_extract_frames, args=args) list(tqdm(p.imap(download_func, saco_yt1b_ids), total=len(saco_yt1b_ids))) done_str = f""" ========== All DONE!! Download, frame extraction, and frame matching is all done! YT1B frames are not ready to use in {args.data_dir}/JPEGImages_6fps Check video frame preparing log at {args.yt1b_frame_prep_log_file} Some videos might not be available any more which will affect the eval reproducibility ========== """ logger.info(done_str) if __name__ == "__main__": main()