| | import os |
| | import json |
| | import copy |
| | import shutil |
| |
|
| | import tqdm |
| |
|
| | import csv |
| |
|
| | TEMPLATE = { |
| | 'set_type': "openset", |
| | 'task': 'bench', |
| | 'data_source': 'dummy', |
| | 'modality': { |
| | "in": [ |
| | "video", |
| | ], |
| | "out": [ |
| | "video" |
| | ] |
| | }, |
| | 'type': 'video comprehension', |
| | 'domain': "Speech", |
| | 'general_capability': ["Reaonsing Ablaity"], |
| | 'version': "1.0", |
| | 'count': 0, |
| | } |
| |
|
| |
|
| |
|
| | METALIST = { |
| | '087_sintel': { |
| | 'task': 'SynVDE', |
| | 'data_source': 'Sintel', |
| | 'modality': { |
| | "in": [ |
| | "video", |
| | ], |
| | "out": [ |
| | "video" |
| | ] |
| | }, |
| | 'domain': "General", |
| | 'general_capability': ["Content Recognition"], |
| | 'version': "1.0", |
| | }, |
| | '088_scannet': { |
| | 'task': 'StaticVDE', |
| | 'data_source': 'ScanNet', |
| | 'modality': { |
| | "in": [ |
| | "video", |
| | ], |
| | "out": [ |
| | "video" |
| | ] |
| | }, |
| | 'domain': "General", |
| | 'general_capability': ["Content Recognition"], |
| | 'version': "1.0", |
| | }, |
| | '089_bonn': { |
| | 'task': 'DynamicVDE', |
| | 'data_source': 'Bonn', |
| | 'modality': { |
| | "in": [ |
| | "video", |
| | ], |
| | "out": [ |
| | "video" |
| | ] |
| | }, |
| | 'domain': "General", |
| | 'general_capability': ["Content Recognition"], |
| | 'version': "1.0", |
| | }, |
| | '090_kitti': { |
| | 'task': 'StreetVDE', |
| | 'data_source': 'KITTI', |
| | 'modality': { |
| | "in": [ |
| | "video", |
| | ], |
| | "out": [ |
| | "video" |
| | ] |
| | }, |
| | 'domain': "General", |
| | 'general_capability': ["Content Recognition"], |
| | 'version': "1.0", |
| | } |
| | } |
| |
|
| |
|
| | import random |
| | open_ratio = 0.4 |
| |
|
| | def deterministic_shuffle(n): |
| | lst = list(range(n)) |
| | random.seed(3407) |
| | random.shuffle(lst) |
| | return lst |
| |
|
| |
|
| | OPEN_SAVE = 'release3/open/' |
| | CLOSE_SAVE = 'release3/close/' |
| |
|
| | if __name__ == '__main__': |
| | for folder in METALIST: |
| | assert os.path.exists(folder), f"Path {folder} does not exist" |
| | |
| | for folder in tqdm.tqdm(METALIST): |
| | |
| | original_list = csv.reader(open(os.path.join(folder, 'split.csv'), newline='')) |
| | original_list = list(original_list)[1:] |
| |
|
| | tot_data = original_list |
| |
|
| | shuffled_indices = deterministic_shuffle(len(tot_data)) |
| | tot_data = [tot_data[idx] for idx in shuffled_indices] |
| |
|
| | open_data_list = tot_data[:int(len(tot_data) * open_ratio)] |
| | closed_data_list = tot_data[int(len(tot_data) * open_ratio):] |
| |
|
| | |
| | open_data_json = copy.deepcopy(TEMPLATE) |
| | open_data_json['set_type'] = "openset" |
| | open_data_json.update(METALIST[folder]) |
| | |
| | open_data_json['count'] = len(open_data_list) |
| | open_save_path = os.path.join(OPEN_SAVE, 'video', 'comprehension', open_data_json['task']) |
| | open_save_video_path = os.path.join(open_save_path, 'video') |
| | open_save_depth_path = os.path.join(open_save_path, 'depth') |
| | os.makedirs(open_save_video_path, exist_ok=True) |
| | os.makedirs(open_save_depth_path, exist_ok=True) |
| | |
| |
|
| | data_list = [] |
| | for item in open_data_list: |
| | input_file, output_file = item |
| | input_file = os.path.basename(input_file) |
| | output_file = os.path.basename(output_file) |
| |
|
| | source_input_path = os.path.join(folder, 'video', input_file) |
| | source_output_path = os.path.join(folder, 'video', output_file) |
| | assert os.path.exists(source_input_path), f"Path {source_input_path} does not exist" |
| | assert os.path.exists(source_output_path), f"Path {source_output_path} does not exist" |
| |
|
| | |
| | new_input_path = os.path.join(open_save_video_path, input_file) |
| | new_output_path = os.path.join(open_save_depth_path, output_file) |
| | shutil.copy(source_input_path, new_input_path) |
| | shutil.copy(source_output_path, new_output_path) |
| | data_list.append({ |
| | 'input': input_file, |
| | 'output': output_file, |
| | 'id': input_file.replace('_rgb_left.mp4', ''), |
| | }) |
| |
|
| | open_data_json['data'] = data_list |
| |
|
| |
|
| | open_json_path = os.path.join(open_save_path, 'annotation.json') |
| | with open(open_json_path, 'w') as f: |
| | json.dump(open_data_json, f, indent=4) |
| | print(f"Open data json saved to {open_json_path}") |
| |
|
| | |
| | closed_data_json = copy.deepcopy(TEMPLATE) |
| | closed_data_json['set_type'] = "closeset" |
| | closed_data_json.update(METALIST[folder]) |
| | closed_data_json['count'] = len(closed_data_list) |
| | closed_save_path = os.path.join(CLOSE_SAVE, 'video', 'comprehension', closed_data_json['task']) |
| | closed_save_video_path = os.path.join(closed_save_path, 'video') |
| | closed_save_depth_path = os.path.join(closed_save_path, 'depth') |
| | os.makedirs(closed_save_video_path, exist_ok=True) |
| | os.makedirs(closed_save_depth_path, exist_ok=True) |
| | data_list = [] |
| |
|
| | for item in closed_data_list: |
| | input_file, output_file = item |
| | input_file = os.path.basename(input_file) |
| | output_file = os.path.basename(output_file) |
| |
|
| | source_input_path = os.path.join(folder, 'video', input_file) |
| | source_output_path = os.path.join(folder, 'video', output_file) |
| | assert os.path.exists(source_input_path), f"Path {source_input_path} does not exist" |
| | assert os.path.exists(source_output_path), f"Path {source_output_path} does not exist" |
| |
|
| | |
| | new_input_path = os.path.join(closed_save_video_path, input_file) |
| | new_output_path = os.path.join(closed_save_depth_path, output_file) |
| | shutil.copy(source_input_path, new_input_path) |
| | shutil.copy(source_output_path, new_output_path) |
| | data_list.append({ |
| | 'input': input_file, |
| | 'output': output_file, |
| | 'id': input_file.replace('_rgb_left.mp4', ''), |
| | }) |
| | closed_data_json['data'] = data_list |
| | closed_json_path = os.path.join(closed_save_path, 'annotation.json') |
| | with open(closed_json_path, 'w') as f: |
| | json.dump(closed_data_json, f, indent=4) |
| | print(f"Closed data json saved to {closed_json_path}") |
| |
|
| |
|