| import torch |
| import torchvision |
| import numpy as np |
| import math |
| import random |
| import time |
|
|
|
|
| class Bucketeer: |
| def __init__( |
| self, dataloader, |
| sizes=[(256, 256), (192, 384), (192, 320), (384, 192), (320, 192)], |
| is_infinite=True, epoch=0, |
| ): |
| |
| self.sizes = sizes |
| self.batch_size = dataloader.batch_size |
| self._dataloader = dataloader |
| self.iterator = iter(dataloader) |
| self.sampler = dataloader.sampler |
| self.buckets = {s: [] for s in self.sizes} |
| self.is_infinite = is_infinite |
| self._epoch = epoch |
|
|
| def get_available_batch(self): |
| available_size = [] |
| for b in self.buckets: |
| if len(self.buckets[b]) >= self.batch_size: |
| available_size.append(b) |
|
|
| if len(available_size) == 0: |
| return None |
| else: |
| b = random.choice(available_size) |
| batch = self.buckets[b][:self.batch_size] |
| self.buckets[b] = self.buckets[b][self.batch_size:] |
| return batch |
|
|
| def __next__(self): |
| batch = self.get_available_batch() |
| while batch is None: |
| try: |
| elements = next(self.iterator) |
| except StopIteration: |
| |
| if self.is_infinite: |
| self._epoch += 1 |
| if hasattr(self._dataloader.sampler, "set_epoch"): |
| self._dataloader.sampler.set_epoch(self._epoch) |
| time.sleep(2) |
| self.iterator = iter(self._dataloader) |
| elements = next(self.iterator) |
| else: |
| raise StopIteration |
|
|
| for dct in elements: |
| try: |
| img = dct['video'] |
| size = (img.shape[-1], img.shape[-2]) |
| self.buckets[size].append({**{'video': img}, **{k:dct[k] for k in dct if k != 'video'}}) |
| except Exception as e: |
| continue |
|
|
| batch = self.get_available_batch() |
|
|
| out = {k:[batch[i][k] for i in range(len(batch))] for k in batch[0]} |
| return {k: torch.stack(o, dim=0) if isinstance(o[0], torch.Tensor) else o for k, o in out.items()} |
|
|
| def __iter__(self): |
| return self |
|
|
| def __len__(self): |
| return len(self.iterator) |
|
|
|
|
| class TemporalLengthBucketeer: |
| def __init__( |
| self, dataloader, max_frames=16, epoch=0, |
| ): |
| self.batch_size = dataloader.batch_size |
| self._dataloader = dataloader |
| self.iterator = iter(dataloader) |
| self.buckets = {temp: [] for temp in range(1, max_frames + 1)} |
| self._epoch = epoch |
|
|
| def get_available_batch(self): |
| available_size = [] |
| for b in self.buckets: |
| if len(self.buckets[b]) >= self.batch_size: |
| available_size.append(b) |
|
|
| if len(available_size) == 0: |
| return None |
| else: |
| b = random.choice(available_size) |
| batch = self.buckets[b][:self.batch_size] |
| self.buckets[b] = self.buckets[b][self.batch_size:] |
| return batch |
|
|
| def __next__(self): |
| batch = self.get_available_batch() |
| while batch is None: |
| try: |
| elements = next(self.iterator) |
| except StopIteration: |
| |
| self._epoch += 1 |
| if hasattr(self._dataloader.sampler, "set_epoch"): |
| self._dataloader.sampler.set_epoch(self._epoch) |
| time.sleep(2) |
| self.iterator = iter(self._dataloader) |
| elements = next(self.iterator) |
|
|
| for dct in elements: |
| try: |
| video_latent = dct['video'] |
| temp = video_latent.shape[2] |
| self.buckets[temp].append({**{'video': video_latent}, **{k:dct[k] for k in dct if k != 'video'}}) |
| except Exception as e: |
| continue |
|
|
| batch = self.get_available_batch() |
|
|
| out = {k:[batch[i][k] for i in range(len(batch))] for k in batch[0]} |
| out = {k: torch.cat(o, dim=0) if isinstance(o[0], torch.Tensor) else o for k, o in out.items()} |
|
|
| if 'prompt_embed' in out: |
| |
| prompt_embeds = out['prompt_embed'].clone() |
| del out['prompt_embed'] |
| prompt_attention_mask = out['prompt_attention_mask'].clone() |
| del out['prompt_attention_mask'] |
| pooled_prompt_embeds = out['pooled_prompt_embed'].clone() |
| del out['pooled_prompt_embed'] |
|
|
| out['text'] = { |
| 'prompt_embeds' : prompt_embeds, |
| 'prompt_attention_mask': prompt_attention_mask, |
| 'pooled_prompt_embeds': pooled_prompt_embeds, |
| } |
|
|
| return out |
|
|
| def __iter__(self): |
| return self |
|
|
| def __len__(self): |
| return len(self.iterator) |