Spaces:
Paused
Paused
| import os | |
| import glob | |
| import torch | |
| import numpy as np | |
| import imageio | |
| import json | |
| import torch.nn.functional as F | |
| import cv2 | |
| def load_blendedmvs_data(basedir): | |
| pose_paths = sorted(glob.glob(os.path.join(basedir, 'pose', '*txt'))) | |
| rgb_paths = sorted(glob.glob(os.path.join(basedir, 'rgb', '*png'))) | |
| all_poses = [] | |
| all_imgs = [] | |
| i_split = [[], []] | |
| for i, (pose_path, rgb_path) in enumerate(zip(pose_paths, rgb_paths)): | |
| i_set = int(os.path.split(rgb_path)[-1][0]) | |
| all_imgs.append((imageio.imread(rgb_path) / 255.).astype(np.float32)) | |
| all_poses.append(np.loadtxt(pose_path).astype(np.float32)) | |
| i_split[i_set].append(i) | |
| imgs = np.stack(all_imgs, 0) | |
| poses = np.stack(all_poses, 0) | |
| i_split.append(i_split[-1]) | |
| path_intrinsics = os.path.join(basedir, 'intrinsics.txt') | |
| H, W = imgs[0].shape[:2] | |
| K = np.loadtxt(path_intrinsics) | |
| focal = float(K[0,0]) | |
| render_poses = torch.Tensor(np.loadtxt(os.path.join(basedir, 'test_traj.txt')).reshape(-1,4,4).astype(np.float32)) | |
| return imgs, poses, render_poses, [H, W, focal], K, i_split | |