| | import logging
|
| | from pathlib import Path
|
| | from typing import Union
|
| |
|
| | import pandas as pd
|
| | import torch
|
| | from tensordict import TensorDict
|
| | from torch.utils.data.dataset import Dataset
|
| |
|
| | from ..utils.dist_utils import local_rank
|
| |
|
| | log = logging.getLogger()
|
| |
|
| |
|
| | class ExtractedVGG(Dataset):
|
| |
|
| | def __init__(
|
| | self,
|
| | tsv_path: Union[str, Path],
|
| | *,
|
| | premade_mmap_dir: Union[str, Path],
|
| | data_dim: dict[str, int],
|
| | ):
|
| | super().__init__()
|
| |
|
| | self.data_dim = data_dim
|
| | self.df_list = pd.read_csv(tsv_path, sep='\t').to_dict('records')
|
| | self.ids = [d['id'] for d in self.df_list]
|
| |
|
| | log.info(f'Loading precomputed mmap from {premade_mmap_dir}')
|
| |
|
| | premade_mmap_dir = Path(premade_mmap_dir)
|
| | td = TensorDict.load_memmap(premade_mmap_dir)
|
| | log.info(f'Loaded precomputed mmap from {premade_mmap_dir}')
|
| | self.mean = td['mean']
|
| | self.std = td['std']
|
| | self.clip_features = td['clip_features']
|
| | self.sync_features = td['sync_features']
|
| | self.text_features = td['text_features']
|
| |
|
| | if local_rank == 0:
|
| | log.info(f'Loaded {len(self)} samples.')
|
| | log.info(f'Loaded mean: {self.mean.shape}.')
|
| | log.info(f'Loaded std: {self.std.shape}.')
|
| | log.info(f'Loaded clip_features: {self.clip_features.shape}.')
|
| | log.info(f'Loaded sync_features: {self.sync_features.shape}.')
|
| | log.info(f'Loaded text_features: {self.text_features.shape}.')
|
| |
|
| | assert self.mean.shape[1] == self.data_dim['latent_seq_len'], \
|
| | f'{self.mean.shape[1]} != {self.data_dim["latent_seq_len"]}'
|
| | assert self.std.shape[1] == self.data_dim['latent_seq_len'], \
|
| | f'{self.std.shape[1]} != {self.data_dim["latent_seq_len"]}'
|
| |
|
| | assert self.clip_features.shape[1] == self.data_dim['clip_seq_len'], \
|
| | f'{self.clip_features.shape[1]} != {self.data_dim["clip_seq_len"]}'
|
| | assert self.sync_features.shape[1] == self.data_dim['sync_seq_len'], \
|
| | f'{self.sync_features.shape[1]} != {self.data_dim["sync_seq_len"]}'
|
| | assert self.text_features.shape[1] == self.data_dim['text_seq_len'], \
|
| | f'{self.text_features.shape[1]} != {self.data_dim["text_seq_len"]}'
|
| |
|
| | assert self.clip_features.shape[-1] == self.data_dim['clip_dim'], \
|
| | f'{self.clip_features.shape[-1]} != {self.data_dim["clip_dim"]}'
|
| | assert self.sync_features.shape[-1] == self.data_dim['sync_dim'], \
|
| | f'{self.sync_features.shape[-1]} != {self.data_dim["sync_dim"]}'
|
| | assert self.text_features.shape[-1] == self.data_dim['text_dim'], \
|
| | f'{self.text_features.shape[-1]} != {self.data_dim["text_dim"]}'
|
| |
|
| | self.video_exist = torch.tensor(1, dtype=torch.bool)
|
| | self.text_exist = torch.tensor(1, dtype=torch.bool)
|
| |
|
| | def compute_latent_stats(self) -> tuple[torch.Tensor, torch.Tensor]:
|
| | latents = self.mean
|
| | return latents.mean(dim=(0, 1)), latents.std(dim=(0, 1))
|
| |
|
| | def get_memory_mapped_tensor(self) -> TensorDict:
|
| | td = TensorDict({
|
| | 'mean': self.mean,
|
| | 'std': self.std,
|
| | 'clip_features': self.clip_features,
|
| | 'sync_features': self.sync_features,
|
| | 'text_features': self.text_features,
|
| | })
|
| | return td
|
| |
|
| | def __getitem__(self, idx: int) -> dict[str, torch.Tensor]:
|
| | data = {
|
| | 'id': self.df_list[idx]['id'],
|
| | 'a_mean': self.mean[idx],
|
| | 'a_std': self.std[idx],
|
| | 'clip_features': self.clip_features[idx],
|
| | 'sync_features': self.sync_features[idx],
|
| | 'text_features': self.text_features[idx],
|
| | 'caption': self.df_list[idx]['label'],
|
| | 'video_exist': self.video_exist,
|
| | 'text_exist': self.text_exist,
|
| | }
|
| |
|
| | return data
|
| |
|
| | def __len__(self):
|
| | return len(self.ids)
|
| |
|