| | from utils.lmdb import get_array_shape_from_lmdb, retrieve_row_from_lmdb |
| | from torch.utils.data import Dataset |
| | import numpy as np |
| | import torch |
| | import lmdb |
| | import json |
| | from pathlib import Path |
| | from PIL import Image |
| | import os |
| |
|
| |
|
| | class TextDataset(Dataset): |
| | def __init__(self, prompt_path, extended_prompt_path=None): |
| | with open(prompt_path, encoding="utf-8") as f: |
| | self.prompt_list = [line.rstrip() for line in f] |
| |
|
| | if extended_prompt_path is not None: |
| | with open(extended_prompt_path, encoding="utf-8") as f: |
| | self.extended_prompt_list = [line.rstrip() for line in f] |
| | assert len(self.extended_prompt_list) == len(self.prompt_list) |
| | else: |
| | self.extended_prompt_list = None |
| |
|
| | def __len__(self): |
| | return len(self.prompt_list) |
| |
|
| | def __getitem__(self, idx): |
| | batch = { |
| | "prompts": self.prompt_list[idx], |
| | "idx": idx, |
| | } |
| | if self.extended_prompt_list is not None: |
| | batch["extended_prompts"] = self.extended_prompt_list[idx] |
| | return batch |
| |
|
| |
|
| | class ODERegressionLMDBDataset(Dataset): |
| | def __init__(self, data_path: str, max_pair: int = int(1e8)): |
| | self.env = lmdb.open(data_path, readonly=True, |
| | lock=False, readahead=False, meminit=False) |
| |
|
| | self.latents_shape = get_array_shape_from_lmdb(self.env, 'latents') |
| | self.max_pair = max_pair |
| |
|
| | def __len__(self): |
| | return min(self.latents_shape[0], self.max_pair) |
| |
|
| | def __getitem__(self, idx): |
| | """ |
| | Outputs: |
| | - prompts: List of Strings |
| | - latents: Tensor of shape (num_denoising_steps, num_frames, num_channels, height, width). It is ordered from pure noise to clean image. |
| | """ |
| | latents = retrieve_row_from_lmdb( |
| | self.env, |
| | "latents", np.float16, idx, shape=self.latents_shape[1:] |
| | ) |
| |
|
| | if len(latents.shape) == 4: |
| | latents = latents[None, ...] |
| |
|
| | prompts = retrieve_row_from_lmdb( |
| | self.env, |
| | "prompts", str, idx |
| | ) |
| | return { |
| | "prompts": prompts, |
| | "ode_latent": torch.tensor(latents, dtype=torch.float32) |
| | } |
| |
|
| |
|
| | class ShardingLMDBDataset(Dataset): |
| | def __init__(self, data_path: str, max_pair: int = int(1e8)): |
| | self.envs = [] |
| | self.index = [] |
| |
|
| | for fname in sorted(os.listdir(data_path)): |
| | path = os.path.join(data_path, fname) |
| | env = lmdb.open(path, |
| | readonly=True, |
| | lock=False, |
| | readahead=False, |
| | meminit=False) |
| | self.envs.append(env) |
| |
|
| | self.latents_shape = [None] * len(self.envs) |
| | for shard_id, env in enumerate(self.envs): |
| | self.latents_shape[shard_id] = get_array_shape_from_lmdb(env, 'latents') |
| | for local_i in range(self.latents_shape[shard_id][0]): |
| | self.index.append((shard_id, local_i)) |
| |
|
| | |
| |
|
| | self.max_pair = max_pair |
| |
|
| | def __len__(self): |
| | return len(self.index) |
| |
|
| | def __getitem__(self, idx): |
| | """ |
| | Outputs: |
| | - prompts: List of Strings |
| | - latents: Tensor of shape (num_denoising_steps, num_frames, num_channels, height, width). It is ordered from pure noise to clean image. |
| | """ |
| | shard_id, local_idx = self.index[idx] |
| |
|
| | latents = retrieve_row_from_lmdb( |
| | self.envs[shard_id], |
| | "latents", np.float16, local_idx, |
| | shape=self.latents_shape[shard_id][1:] |
| | ) |
| |
|
| | if len(latents.shape) == 4: |
| | latents = latents[None, ...] |
| |
|
| | prompts = retrieve_row_from_lmdb( |
| | self.envs[shard_id], |
| | "prompts", str, local_idx |
| | ) |
| |
|
| | return { |
| | "prompts": prompts, |
| | "ode_latent": torch.tensor(latents, dtype=torch.float32) |
| | } |
| |
|
| |
|
| | class TextImagePairDataset(Dataset): |
| | def __init__( |
| | self, |
| | data_dir, |
| | transform=None, |
| | eval_first_n=-1, |
| | pad_to_multiple_of=None |
| | ): |
| | """ |
| | Args: |
| | data_dir (str): Path to the directory containing: |
| | - target_crop_info_*.json (metadata file) |
| | - */ (subdirectory containing images with matching aspect ratio) |
| | transform (callable, optional): Optional transform to be applied on the image |
| | """ |
| | self.transform = transform |
| | data_dir = Path(data_dir) |
| |
|
| | |
| | metadata_files = list(data_dir.glob('target_crop_info_*.json')) |
| | if not metadata_files: |
| | raise FileNotFoundError(f"No metadata file found in {data_dir}") |
| | if len(metadata_files) > 1: |
| | raise ValueError(f"Multiple metadata files found in {data_dir}") |
| |
|
| | metadata_path = metadata_files[0] |
| | |
| | aspect_ratio = metadata_path.stem.split('_')[-1] |
| |
|
| | |
| | self.image_dir = data_dir / aspect_ratio |
| | if not self.image_dir.exists(): |
| | raise FileNotFoundError(f"Image directory not found: {self.image_dir}") |
| |
|
| | |
| | with open(metadata_path, 'r') as f: |
| | self.metadata = json.load(f) |
| |
|
| | eval_first_n = eval_first_n if eval_first_n != -1 else len(self.metadata) |
| | self.metadata = self.metadata[:eval_first_n] |
| |
|
| | |
| | for item in self.metadata: |
| | image_path = self.image_dir / item['file_name'] |
| | if not image_path.exists(): |
| | raise FileNotFoundError(f"Image not found: {image_path}") |
| |
|
| | self.dummy_prompt = "DUMMY PROMPT" |
| | self.pre_pad_len = len(self.metadata) |
| | if pad_to_multiple_of is not None and len(self.metadata) % pad_to_multiple_of != 0: |
| | |
| | self.metadata += [self.metadata[-1]] * ( |
| | pad_to_multiple_of - len(self.metadata) % pad_to_multiple_of |
| | ) |
| |
|
| | def __len__(self): |
| | return len(self.metadata) |
| |
|
| | def __getitem__(self, idx): |
| | """ |
| | Returns: |
| | dict: A dictionary containing: |
| | - image: PIL Image |
| | - caption: str |
| | - target_bbox: list of int [x1, y1, x2, y2] |
| | - target_ratio: str |
| | - type: str |
| | - origin_size: tuple of int (width, height) |
| | """ |
| | item = self.metadata[idx] |
| |
|
| | |
| | image_path = self.image_dir / item['file_name'] |
| | image = Image.open(image_path).convert('RGB') |
| |
|
| | |
| | if self.transform: |
| | image = self.transform(image) |
| |
|
| | return { |
| | 'image': image, |
| | 'prompts': item['caption'], |
| | 'target_bbox': item['target_crop']['target_bbox'], |
| | 'target_ratio': item['target_crop']['target_ratio'], |
| | 'type': item['type'], |
| | 'origin_size': (item['origin_width'], item['origin_height']), |
| | 'idx': idx |
| | } |
| |
|
| |
|
| | def cycle(dl): |
| | while True: |
| | for data in dl: |
| | yield data |
| |
|