| | |
| | import argparse |
| | import binascii |
| | import logging |
| | import os |
| | import os.path as osp |
| | import shutil |
| | import subprocess |
| |
|
| | import imageio |
| | import torch |
| | import torchvision |
| |
|
| | __all__ = ['save_video', 'save_image', 'str2bool'] |
| |
|
| |
|
| | def rand_name(length=8, suffix=''): |
| | name = binascii.b2a_hex(os.urandom(length)).decode('utf-8') |
| | if suffix: |
| | if not suffix.startswith('.'): |
| | suffix = '.' + suffix |
| | name += suffix |
| | return name |
| |
|
| |
|
| | def merge_video_audio(video_path: str, audio_path: str): |
| | """ |
| | Merge the video and audio into a new video, with the duration set to the shorter of the two, |
| | and overwrite the original video file. |
| | |
| | Parameters: |
| | video_path (str): Path to the original video file |
| | audio_path (str): Path to the audio file |
| | """ |
| | |
| | logging.basicConfig(level=logging.INFO) |
| |
|
| | |
| | if not os.path.exists(video_path): |
| | raise FileNotFoundError(f"video file {video_path} does not exist") |
| | if not os.path.exists(audio_path): |
| | raise FileNotFoundError(f"audio file {audio_path} does not exist") |
| |
|
| | base, ext = os.path.splitext(video_path) |
| | temp_output = f"{base}_temp{ext}" |
| |
|
| | try: |
| | |
| | command = [ |
| | 'ffmpeg', |
| | '-y', |
| | '-i', |
| | video_path, |
| | '-i', |
| | audio_path, |
| | '-c:v', |
| | 'copy', |
| | '-c:a', |
| | 'aac', |
| | '-b:a', |
| | '192k', |
| | '-map', |
| | '0:v:0', |
| | '-map', |
| | '1:a:0', |
| | '-shortest', |
| | temp_output |
| | ] |
| |
|
| | |
| | logging.info("Start merging video and audio...") |
| | result = subprocess.run( |
| | command, stdout=subprocess.PIPE, stderr=subprocess.PIPE, text=True) |
| |
|
| | |
| | if result.returncode != 0: |
| | error_msg = f"FFmpeg execute failed: {result.stderr}" |
| | logging.error(error_msg) |
| | raise RuntimeError(error_msg) |
| |
|
| | shutil.move(temp_output, video_path) |
| | logging.info(f"Merge completed, saved to {video_path}") |
| |
|
| | except Exception as e: |
| | if os.path.exists(temp_output): |
| | os.remove(temp_output) |
| | logging.error(f"merge_video_audio failed with error: {e}") |
| |
|
| |
|
| | def save_video(tensor, |
| | save_file=None, |
| | fps=30, |
| | suffix='.mp4', |
| | nrow=8, |
| | normalize=True, |
| | value_range=(-1, 1)): |
| | |
| | cache_file = osp.join('/tmp', rand_name( |
| | suffix=suffix)) if save_file is None else save_file |
| |
|
| | |
| | try: |
| | |
| | tensor = tensor.clamp(min(value_range), max(value_range)) |
| | tensor = torch.stack([ |
| | torchvision.utils.make_grid( |
| | u, nrow=nrow, normalize=normalize, value_range=value_range) |
| | for u in tensor.unbind(2) |
| | ], |
| | dim=1).permute(1, 2, 3, 0) |
| | tensor = (tensor * 255).type(torch.uint8).cpu() |
| |
|
| | |
| | writer = imageio.get_writer( |
| | cache_file, fps=fps, codec='libx264', quality=8) |
| | for frame in tensor.numpy(): |
| | writer.append_data(frame) |
| | writer.close() |
| | except Exception as e: |
| | logging.info(f'save_video failed, error: {e}') |
| |
|
| |
|
| | def save_image(tensor, save_file, nrow=8, normalize=True, value_range=(-1, 1)): |
| | |
| | suffix = osp.splitext(save_file)[1] |
| | if suffix.lower() not in [ |
| | '.jpg', '.jpeg', '.png', '.tiff', '.gif', '.webp' |
| | ]: |
| | suffix = '.png' |
| |
|
| | |
| | try: |
| | tensor = tensor.clamp(min(value_range), max(value_range)) |
| | torchvision.utils.save_image( |
| | tensor, |
| | save_file, |
| | nrow=nrow, |
| | normalize=normalize, |
| | value_range=value_range) |
| | return save_file |
| | except Exception as e: |
| | logging.info(f'save_image failed, error: {e}') |
| |
|
| |
|
| | def str2bool(v): |
| | """ |
| | Convert a string to a boolean. |
| | |
| | Supported true values: 'yes', 'true', 't', 'y', '1' |
| | Supported false values: 'no', 'false', 'f', 'n', '0' |
| | |
| | Args: |
| | v (str): String to convert. |
| | |
| | Returns: |
| | bool: Converted boolean value. |
| | |
| | Raises: |
| | argparse.ArgumentTypeError: If the value cannot be converted to boolean. |
| | """ |
| | if isinstance(v, bool): |
| | return v |
| | v_lower = v.lower() |
| | if v_lower in ('yes', 'true', 't', 'y', '1'): |
| | return True |
| | elif v_lower in ('no', 'false', 'f', 'n', '0'): |
| | return False |
| | else: |
| | raise argparse.ArgumentTypeError('Boolean value expected (True/False)') |
| |
|
| |
|
| | def masks_like(tensor, zero=False, generator=None, p=0.2): |
| | assert isinstance(tensor, list) |
| | out1 = [torch.ones(u.shape, dtype=u.dtype, device=u.device) for u in tensor] |
| |
|
| | out2 = [torch.ones(u.shape, dtype=u.dtype, device=u.device) for u in tensor] |
| |
|
| | if zero: |
| | if generator is not None: |
| | for u, v in zip(out1, out2): |
| | random_num = torch.rand( |
| | 1, generator=generator, device=generator.device).item() |
| | if random_num < p: |
| | u[:, 0] = torch.normal( |
| | mean=-3.5, |
| | std=0.5, |
| | size=(1,), |
| | device=u.device, |
| | generator=generator).expand_as(u[:, 0]).exp() |
| | v[:, 0] = torch.zeros_like(v[:, 0]) |
| | else: |
| | u[:, 0] = u[:, 0] |
| | v[:, 0] = v[:, 0] |
| | else: |
| | for u, v in zip(out1, out2): |
| | u[:, 0] = torch.zeros_like(u[:, 0]) |
| | v[:, 0] = torch.zeros_like(v[:, 0]) |
| |
|
| | return out1, out2 |
| |
|
| |
|
| | def best_output_size(w, h, dw, dh, expected_area): |
| | |
| | ratio = w / h |
| | ow = (expected_area * ratio)**0.5 |
| | oh = expected_area / ow |
| |
|
| | |
| | ow1 = int(ow // dw * dw) |
| | oh1 = int(expected_area / ow1 // dh * dh) |
| | assert ow1 % dw == 0 and oh1 % dh == 0 and ow1 * oh1 <= expected_area |
| | ratio1 = ow1 / oh1 |
| |
|
| | |
| | oh2 = int(oh // dh * dh) |
| | ow2 = int(expected_area / oh2 // dw * dw) |
| | assert oh2 % dh == 0 and ow2 % dw == 0 and ow2 * oh2 <= expected_area |
| | ratio2 = ow2 / oh2 |
| |
|
| | |
| | if max(ratio / ratio1, ratio1 / ratio) < max(ratio / ratio2, |
| | ratio2 / ratio): |
| | return ow1, oh1 |
| | else: |
| | return ow2, oh2 |
| |
|