|
|
|
|
|
import argparse |
|
|
import logging |
|
|
|
|
|
import multiprocessing as mp |
|
|
import os |
|
|
from functools import partial |
|
|
|
|
|
import pandas as pd |
|
|
from saco_yt1b_frame_prep_util import YtVideoPrep |
|
|
from tqdm import tqdm |
|
|
|
|
|
logger = logging.getLogger(__name__) |
|
|
|
|
|
|
|
|
def download_and_extract_frames(saco_yt1b_id, args): |
|
|
video_prep = YtVideoPrep( |
|
|
saco_yt1b_id=saco_yt1b_id, |
|
|
data_dir=args.data_dir, |
|
|
cookies_file=args.cookies_file, |
|
|
yt1b_start_end_time_file=args.yt1b_start_end_time_file, |
|
|
ffmpeg_timeout=args.ffmpeg_timeout, |
|
|
sleep_interval=args.sleep_interval, |
|
|
max_sleep_interval=args.max_sleep_interval, |
|
|
) |
|
|
|
|
|
status = video_prep.download_youtube_video() |
|
|
logger.info(f"[video download][{saco_yt1b_id}] download status {status}") |
|
|
|
|
|
if status not in ["already exists", "success"]: |
|
|
logger.warning( |
|
|
f"Video download failed for {saco_yt1b_id}, skipping frame generation" |
|
|
) |
|
|
return False |
|
|
|
|
|
status = video_prep.extract_frames_in_6fps_and_width_1080() |
|
|
logger.info(f"[frame extracting][{saco_yt1b_id}] frame extracting status {status}") |
|
|
return True |
|
|
|
|
|
|
|
|
def main(): |
|
|
parser = argparse.ArgumentParser() |
|
|
parser.add_argument( |
|
|
"--data_dir", |
|
|
type=str, |
|
|
required=True, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--cookies_file", |
|
|
type=str, |
|
|
required=True, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--yt1b_start_end_time_file", |
|
|
type=str, |
|
|
required=True, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--yt1b_frame_prep_log_file", |
|
|
type=str, |
|
|
required=True, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--ffmpeg_timeout", |
|
|
type=str, |
|
|
default=7200, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--sleep_interval", |
|
|
type=int, |
|
|
default=10, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--max_sleep_interval", |
|
|
type=int, |
|
|
default=30, |
|
|
) |
|
|
parser.add_argument( |
|
|
"--num_workers", |
|
|
type=int, |
|
|
default=4, |
|
|
) |
|
|
args = parser.parse_args() |
|
|
|
|
|
log_dir = os.path.dirname(args.yt1b_frame_prep_log_file) |
|
|
if log_dir: |
|
|
os.makedirs(log_dir, exist_ok=True) |
|
|
|
|
|
|
|
|
|
|
|
logging.basicConfig( |
|
|
level=logging.INFO, |
|
|
format="%(asctime)s [%(processName)s/%(threadName)s] %(name)s - %(levelname)s: %(message)s", |
|
|
handlers=[ |
|
|
logging.FileHandler(args.yt1b_frame_prep_log_file, mode="w"), |
|
|
logging.StreamHandler(), |
|
|
], |
|
|
force=True, |
|
|
) |
|
|
|
|
|
YT_DLP_WARNING_STR = """ ========== |
|
|
NOTICE!! |
|
|
This script uses yt-dlp to download youtube videos. |
|
|
See the youtube account banning risk in https://github.com/yt-dlp/yt-dlp/wiki/Extractors#exporting-youtube-cookies |
|
|
========== |
|
|
""" |
|
|
|
|
|
logger.info(YT_DLP_WARNING_STR) |
|
|
|
|
|
args = parser.parse_args() |
|
|
|
|
|
with open(args.yt1b_start_end_time_file, "r") as f: |
|
|
yt1b_start_end_time_df = pd.read_json(f) |
|
|
|
|
|
saco_yt1b_ids = yt1b_start_end_time_df.saco_yt1b_id.unique() |
|
|
num_workers = args.num_workers |
|
|
logger.info( |
|
|
f"Starting with {num_workers} parallel worker(s) (sleep_interval={args.sleep_interval}-{args.max_sleep_interval}s)" |
|
|
) |
|
|
|
|
|
with mp.Pool(num_workers) as p: |
|
|
download_func = partial(download_and_extract_frames, args=args) |
|
|
list(tqdm(p.imap(download_func, saco_yt1b_ids), total=len(saco_yt1b_ids))) |
|
|
|
|
|
done_str = f""" ========== |
|
|
All DONE!! |
|
|
Download, frame extraction, and frame matching is all done! YT1B frames are not ready to use in {args.data_dir}/JPEGImages_6fps |
|
|
Check video frame preparing log at {args.yt1b_frame_prep_log_file} |
|
|
Some videos might not be available any more which will affect the eval reproducibility |
|
|
========== |
|
|
""" |
|
|
logger.info(done_str) |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|