| import os |
| import pickle |
| import random |
| import numpy as np |
| import pandas as pd |
| from collections import defaultdict |
|
|
| import torch |
| import torchvision |
| from torch.utils.data import DataLoader, Dataset, Sampler |
|
|
| from video_reader import PyVideoReader |
|
|
| from diffusers.utils import export_to_video |
| from diffusers.training_utils import free_memory |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| resolution_bucket_options = { |
| 640: [ |
| (768, 320), |
| (768, 384), |
| (640, 384), |
| (768, 512), |
| (576, 448), |
| (512, 512), |
| (448, 576), |
| (512, 768), |
| (384, 640), |
| (384, 768), |
| (320, 768), |
| ], |
| } |
|
|
| length_bucket_options = { |
| 1: [321, 301, 281, 261, 241, 221, 193, 181, 161, 141, 121, 101, 81, 61, 41, 21], |
| 2: [193, 177, 161, 156, 145, 133, 129, 121, 113, 109, 97, 85, 81, 73, 65, 61, 49, 37, 25], |
| } |
|
|
| def find_nearest_resolution_bucket(h, w, resolution=640): |
| min_metric = float('inf') |
| best_bucket = None |
| for (bucket_h, bucket_w) in resolution_bucket_options[resolution]: |
| metric = abs(h * bucket_w - w * bucket_h) |
| if metric <= min_metric: |
| min_metric = metric |
| best_bucket = (bucket_h, bucket_w) |
| return best_bucket |
|
|
| def find_nearest_length_bucket(length, stride=1): |
| buckets = length_bucket_options[stride] |
| min_bucket = min(buckets) |
| if length < min_bucket: |
| return length |
| valid_buckets = [bucket for bucket in buckets if bucket <= length] |
| return max(valid_buckets) |
|
|
| def read_cut_crop_and_resize(video_path, f_prime, h_prime, w_prime, stride=1): |
| vr = PyVideoReader(video_path, threads=0) |
| total_frames = len(vr) |
|
|
| if stride != 1: |
| required_span = stride * (f_prime - 1) |
| start_frame = max(0, total_frames - required_span - 1) |
| else: |
| start_frame = max(0, total_frames - f_prime) |
|
|
| frame_indices = list(range(start_frame, total_frames, stride)) |
| assert len(frame_indices) == f_prime |
| frames = torch.from_numpy(vr.get_batch(frame_indices)).float() |
|
|
|
|
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
|
|
| |
| |
| |
| |
| |
|
|
|
|
| frames = (frames / 127.5) - 1 |
| video = frames.permute(0, 3, 1, 2) |
|
|
| frames, channels, h, w = video.shape |
| aspect_ratio_original = h / w |
| aspect_ratio_target = h_prime / w_prime |
|
|
| if aspect_ratio_original >= aspect_ratio_target: |
| new_h = int(w * aspect_ratio_target) |
| top = (h - new_h) // 2 |
| bottom = top + new_h |
| left = 0 |
| right = w |
| else: |
| new_w = int(h / aspect_ratio_target) |
| left = (w - new_w) // 2 |
| right = left + new_w |
| top = 0 |
| bottom = h |
|
|
| |
| cropped_video = video[:, :, top:bottom, left:right] |
| |
| resized_video = torchvision.transforms.functional.resize(cropped_video, (h_prime, w_prime)) |
| return resized_video |
|
|
| def save_frames(frame_raw, fps=24, video_path="1.mp4"): |
| save_list = [] |
| for frame in frame_raw: |
| frame = (frame + 1) / 2 * 255 |
| frame = torchvision.transforms.transforms.ToPILImage()(frame.to(torch.uint8)).convert("RGB") |
| save_list.append(frame) |
| frame = None |
| del frame |
| export_to_video(save_list, video_path, fps=fps) |
| |
| save_list = None |
| del save_list |
| free_memory() |
|
|
| class BucketedFeatureDataset(Dataset): |
| def __init__(self, csv_file, video_folder, stride=1, cache_file=None, force_rebuild=False): |
| self.csv_file = csv_file |
| self.video_folder = video_folder |
| self.stride = stride |
| |
| if cache_file is None: |
| cache_file = os.path.join(video_folder, f"dataset_cache_stride{stride}.pkl") |
|
|
| if force_rebuild or not os.path.exists(cache_file): |
| print("Building metadata cache...") |
| self._build_metadata() |
| self._save_cache(cache_file) |
| else: |
| print("Loading cached metadata...") |
| with open(cache_file, "rb") as f: |
| cached_data = pickle.load(f) |
| if cached_data.get("stride", 1) != stride: |
| print(f"Stride mismatch in cache (cached: {cached_data.get('stride', 1)}, current: {stride}). Rebuilding...") |
| self._build_metadata() |
| self._save_cache(cache_file) |
| else: |
| self.samples = cached_data["samples"] |
| self.buckets = cached_data["buckets"] |
| print(f"Loaded {len(self.samples)} samples from cache") |
| |
|
|
| def _save_cache(self, cache_file): |
| print("Saving metadata cache...") |
| cached_data = { |
| "samples": self.samples, |
| "buckets": self.buckets, |
| "stride": self.stride |
| } |
| with open(cache_file, "wb") as f: |
| pickle.dump(cached_data, f) |
| print(f"Cached {len(self.samples)} samples with stride={self.stride}") |
|
|
| |
| |
| |
| |
| |
|
|
| |
| |
| |
| |
|
|
| |
|
|
| |
| |
| |
| |
| |
| |
|
|
| def _build_metadata(self): |
| self.df = pd.read_csv(self.csv_file) |
| |
| self.samples = [] |
| self.buckets = defaultdict(list) |
| sample_idx = 0 |
| |
| print(f"Processing {len(self.df)} records from CSV with stride={self.stride}...") |
| for i, row in self.df.iterrows(): |
| if i % 10000 == 0: |
| print(f"Processed {i}/{len(self.df)} records") |
| |
| uttid = os.path.basename(row['videoFile']).replace(".mp4", "") |
| video_file = row['videoFile'] |
| video_path = os.path.join(self.video_folder, video_file) |
| prompt = row["caption"] |
| num_frame = row["num_frame"] |
| height = row["height"] |
| width = row["width"] |
| fps = row["fps"] |
|
|
| |
| |
| |
|
|
| effective_num_frame = (num_frame + self.stride - 1) // self.stride |
| bucket_height, bucket_width = find_nearest_resolution_bucket(height, width, resolution=640) |
| bucket_num_frame = find_nearest_length_bucket(effective_num_frame, stride=self.stride) |
| bucket_key = (bucket_num_frame, bucket_height, bucket_width) |
|
|
| sample_info = { |
| "uttid": uttid, |
| "bucket_key": bucket_key, |
| "video_path": video_path, |
| "prompt": prompt, |
| "fps": fps, |
| "stride": self.stride, |
| "effective_num_frame": effective_num_frame, |
| "num_frame": num_frame, |
| "height": height, |
| "width": width, |
| "bucket_num_frame": bucket_num_frame, |
| "bucket_height": bucket_height, |
| "bucket_width": bucket_width, |
| } |
|
|
| self.samples.append(sample_info) |
| self.buckets[bucket_key].append(sample_idx) |
| sample_idx += 1 |
|
|
| def __len__(self): |
| return len(self.samples) |
|
|
| def __getitem__(self, idx): |
| |
| |
| |
| |
| |
| |
| |
| |
| while True: |
| sample_info = self.samples[idx] |
| try: |
| video_data = read_cut_crop_and_resize( |
| video_path=sample_info["video_path"], |
| f_prime=sample_info["bucket_num_frame"], |
| h_prime=sample_info["bucket_height"], |
| w_prime=sample_info["bucket_width"], |
| stride=self.stride, |
| ) |
| break |
| except Exception: |
| idx = random.randint(0, len(self.samples) - 1) |
| print(f"Error loading {sample_info['video_path']}, retrying...") |
|
|
| return { |
| "uttid": sample_info["uttid"], |
| "bucket_key": sample_info["bucket_key"], |
| "video_metadata": { |
| "num_frames": sample_info["bucket_num_frame"], |
| "height": sample_info["bucket_height"], |
| "width": sample_info["bucket_width"], |
| "fps": sample_info["fps"], |
| "stride": self.stride, |
| "effective_num_frame": sample_info["effective_num_frame"], |
| }, |
| "videos": video_data, |
| "prompts": sample_info["prompt"], |
| "first_frames_images": (video_data[0] + 1) / 2 * 255, |
| } |
|
|
| class BucketedSampler(Sampler): |
| def __init__(self, dataset, batch_size, drop_last=False, shuffle=False, seed=42): |
| self.dataset = dataset |
| self.batch_size = batch_size |
| self.drop_last = drop_last |
| self.shuffle = shuffle |
| self.seed = seed |
| self.generator = torch.Generator() |
| self.buckets = dataset.buckets |
| self._epoch = 0 |
|
|
| def set_epoch(self, epoch): |
| self._epoch = epoch |
|
|
| def __iter__(self): |
| if self.shuffle: |
| self.generator.manual_seed(self.seed + self._epoch) |
| else: |
| self.generator.manual_seed(self.seed) |
|
|
| bucket_iterators = {} |
| bucket_batches = {} |
|
|
| for bucket_key, sample_indices in self.buckets.items(): |
| indices = sample_indices.copy() |
| if self.shuffle: |
| indices = torch.randperm(len(indices), generator=self.generator).tolist() |
| indices = [sample_indices[i] for i in indices] |
|
|
| batches = [] |
| for i in range(0, len(indices), self.batch_size): |
| batch = indices[i : i + self.batch_size] |
| if len(batch) == self.batch_size or not self.drop_last: |
| batches.append(batch) |
|
|
| if batches: |
| bucket_batches[bucket_key] = batches |
| bucket_iterators[bucket_key] = iter(batches) |
|
|
| remaining_buckets = list(bucket_iterators.keys()) |
|
|
| while remaining_buckets: |
| idx = torch.randint(len(remaining_buckets), (1,), generator=self.generator).item() |
| bucket_key = remaining_buckets[idx] |
|
|
| bucket_iter = bucket_iterators[bucket_key] |
|
|
| try: |
| batch = next(bucket_iter) |
| yield batch |
| except StopIteration: |
| remaining_buckets.remove(bucket_key) |
|
|
| def __len__(self): |
| total_batches = 0 |
| for sample_indices in self.buckets.values(): |
| num_batches = len(sample_indices) // self.batch_size |
| if not self.drop_last and len(sample_indices) % self.batch_size != 0: |
| num_batches += 1 |
| total_batches += num_batches |
| return total_batches |
|
|
|
|
| def collate_fn(batch): |
| def collate_dict(data_list): |
| if isinstance(data_list[0], dict): |
| return { |
| key: collate_dict([d[key] for d in data_list]) |
| for key in data_list[0] |
| } |
| elif isinstance(data_list[0], torch.Tensor): |
| return torch.stack(data_list) |
| else: |
| return data_list |
| |
| return { |
| key: collate_dict([d[key] for d in batch]) |
| for key in batch[0] |
| } |
|
|
|
|
| if __name__ == "__main__": |
| from accelerate import Accelerator |
|
|
| base_name = "sekai-game-drone" |
| csv_file = f"/mnt/bn/yufan-dev-my/ysh/Ckpts/Lixsp11/0_final_sekai_dataset/yamls/{base_name}_updated.csv" |
| video_folder = f"/mnt/bn/yufan-dev-my/ysh/Ckpts/Lixsp11/0_final_sekai_dataset/{base_name}" |
| stride = 1 |
| batch_size = 2 |
| num_train_epochs = 1 |
| seed = 0 |
| output_dir = "accelerate_checkpoints" |
| checkpoint_dirs = ( |
| [ |
| d |
| for d in os.listdir(output_dir) |
| if d.startswith("checkpoint-") and os.path.isdir(os.path.join(output_dir, d)) |
| ] |
| if os.path.exists(output_dir) |
| else [] |
| ) |
|
|
| dataset = BucketedFeatureDataset(csv_file=csv_file, video_folder=video_folder, stride=stride) |
| sampler = BucketedSampler(dataset, batch_size=2, drop_last=False, shuffle=True, seed=seed) |
| dataloader = DataLoader(dataset, batch_sampler=sampler, collate_fn=collate_fn, num_workers=8) |
|
|
| print(len(dataset), len(dataloader)) |
| accelerator = Accelerator() |
| dataloader = accelerator.prepare(dataloader) |
| print(f"Dataset size: {len(dataset)}, Dataloader batches: {len(dataloader)}") |
| print(f"Process index: {accelerator.process_index}, World size: {accelerator.num_processes}") |
|
|
| step = 0 |
| global_step = 0 |
| first_epoch = 0 |
| num_update_steps_per_epoch = len(dataloader) |
|
|
| print("Testing dataloader...") |
| step = global_step |
| for epoch in range(first_epoch, num_train_epochs): |
| sampler.set_epoch(epoch) |
| skip_steps = 0 |
| printed_skip_log = False |
| for i, batch in enumerate(dataloader): |
| if epoch == first_epoch and skip_steps < (global_step % num_update_steps_per_epoch): |
| skip_steps += 1 |
| continue |
| if epoch == first_epoch and not printed_skip_log: |
| print(f"Skip {skip_steps} steps in epoch {epoch}") |
| printed_skip_log = True |
|
|
| |
| uttid = batch["uttid"] |
| bucket_key = batch["bucket_key"] |
| num_frame = batch["video_metadata"]["num_frames"] |
| height = batch["video_metadata"]["height"] |
| width = batch["video_metadata"]["width"] |
|
|
| |
| video_data = batch["videos"] |
| prompt = batch["prompts"] |
| first_frames_images = batch["first_frames_images"] |
| first_frames_images = [torchvision.transforms.ToPILImage()(x.to(torch.uint8)) for x in first_frames_images] |
|
|
| |
| |
|
|
| if accelerator.process_index == 0: |
| |
| print(f" Step {step}:") |
| print(f" Batch {i}:") |
| print(f" Batch size: {len(uttid)}") |
| print(f" Uttids: {uttid}") |
| print(f" Dimensions - frames: {num_frame[0]}, height: {height[0]}, width: {width[0]}") |
| print(f" Bucket key: {bucket_key[0]}") |
| print(f" Videos shape: {video_data.shape}") |
| print(f" Cpation: {prompt}") |
|
|
| |
| assert all(nf == num_frame[0] for nf in num_frame), "Frame numbers not consistent in batch" |
| assert all(h == height[0] for h in height), "Heights not consistent in batch" |
| assert all(w == width[0] for w in width), "Widths not consistent in batch" |
|
|
| print(" ✓ Batch dimensions are consistent") |
|
|
| step += 1 |