|
|
---
|
|
|
license: cc-by-4.0
|
|
|
---
|
|
|
|
|
|
# Neural 3D Video Dataset - Processed
|
|
|
|
|
|
This directory contains preprocessed multi-view video data from the Neural 3D Video dataset, converted into a format suitable for 4D reconstruction and novel view synthesis tasks.
|
|
|
|
|
|
## Dataset Overview
|
|
|
|
|
|
**Source Dataset**: Neural 3D Video
|
|
|
**License**: CC-BY-4.0
|
|
|
**Processed Scenes**: 5 dynamic cooking scenes captured from multiple camera angles
|
|
|
|
|
|
### Scenes
|
|
|
|
|
|
| Scene | Description | Cameras | Frames |
|
|
|
|-------|-------------|---------|--------|
|
|
|
| `coffee_martini` | Making a coffee martini cocktail | 18 | 32 |
|
|
|
| `cook_spinach` | Cooking spinach in a pan | 18 | 32 |
|
|
|
| `cut_roasted_beef` | Cutting roasted beef | 18 | 32 |
|
|
|
| `flame_salmon_1` | Flambé salmon preparation | 18 | 32 |
|
|
|
| `sear_steak` | Searing steak in a pan | 18 | 32 |
|
|
|
|
|
|
## Directory Structure
|
|
|
|
|
|
```
|
|
|
Neural-3D-Video-Dataset/
|
|
|
├── README.md (this file)
|
|
|
├── coffee_martini_processed/
|
|
|
│ ├── 256/ # 256×256 resolution
|
|
|
│ │ ├── images/ # 32 frame images
|
|
|
│ │ │ ├── sample_000_cam00.jpg
|
|
|
│ │ │ ├── sample_001_cam01.jpg
|
|
|
│ │ │ └── ...
|
|
|
│ │ ├── transforms.json # Camera poses (JSON format)
|
|
|
│ │ ├── transforms.npz # Camera poses (NumPy format)
|
|
|
│ │ └── camera_visualization.html # Interactive 3D camera viewer
|
|
|
│ └── 512/ # 512×512 resolution
|
|
|
│ ├── images/
|
|
|
│ ├── transforms.json
|
|
|
│ ├── transforms.npz
|
|
|
│ └── camera_visualization.html
|
|
|
├── cook_spinach_processed/
|
|
|
│ ├── 256/ ...
|
|
|
│ └── 512/ ...
|
|
|
├── cut_roasted_beef_processed/
|
|
|
│ ├── 256/ ...
|
|
|
│ └── 512/ ...
|
|
|
├── flame_salmon_1_processed/
|
|
|
│ ├── 256/ ...
|
|
|
│ └── 512/ ...
|
|
|
└── sear_steak_processed/
|
|
|
├── 256/ ...
|
|
|
└── 512/ ...
|
|
|
```
|
|
|
|
|
|
## Data Format
|
|
|
|
|
|
### Camera Poses (`transforms.json`)
|
|
|
|
|
|
The camera poses are stored in a JSON file with the following structure:
|
|
|
|
|
|
```json
|
|
|
{
|
|
|
"frames": [
|
|
|
{
|
|
|
"front": {
|
|
|
"timestamp": 0,
|
|
|
"file_path": "./images/sample_000_cam00.jpg",
|
|
|
"w": 256,
|
|
|
"h": 256,
|
|
|
"fx": 341.33,
|
|
|
"fy": 341.33,
|
|
|
"cx": 128.0,
|
|
|
"cy": 128.0,
|
|
|
"w2c": [[...], [...], [...], [...]], // 4×4 world-to-camera matrix
|
|
|
"c2w": [[...], [...], [...]], // 3×4 camera-to-world matrix
|
|
|
"blender_camera_location": [x, y, z] // Camera position in world coordinates
|
|
|
}
|
|
|
},
|
|
|
...
|
|
|
]
|
|
|
}
|
|
|
```
|
|
|
|
|
|
**Intrinsics** (camera internal parameters):
|
|
|
- **256×256**: `fx = fy = 341.33`, `cx = cy = 128.0`
|
|
|
- **512×512**: `fx = fy = 682.67`, `cx = cy = 256.0`
|
|
|
|
|
|
**Extrinsics** (camera external parameters):
|
|
|
- `w2c`: 4×4 world-to-camera transformation matrix
|
|
|
- `c2w`: 3×4 camera-to-world transformation matrix (rotation + translation)
|
|
|
- `blender_camera_location`: 3D camera position `[x, y, z]` in world coordinates
|
|
|
|
|
|
### NumPy Format (`transforms.npz`)
|
|
|
|
|
|
For convenience, camera parameters are also provided in NumPy format:
|
|
|
|
|
|
```python
|
|
|
import numpy as np
|
|
|
|
|
|
data = np.load('transforms.npz')
|
|
|
intrinsics = data['intrinsics'] # (32, 3, 3) - intrinsic matrices
|
|
|
extrinsics_w2c = data['extrinsics_w2c'] # (32, 4, 4) - world-to-camera
|
|
|
extrinsics_c2w = data['extrinsics_c2w'] # (32, 4, 4) - camera-to-world
|
|
|
camera_positions = data['camera_positions'] # (32, 3) - camera locations
|
|
|
```
|
|
|
|
|
|
### Frame Images
|
|
|
|
|
|
- **Format**: JPEG
|
|
|
- **Resolutions**: 256×256 and 512×512
|
|
|
- **Count**: 32 frames per scene
|
|
|
- **Naming**: `sample_{frame:03d}_cam{camera:02d}.jpg`
|
|
|
|
|
|
Each frame is extracted from a different camera view:
|
|
|
- Frame 0 → cam00
|
|
|
- Frame 1 → cam01
|
|
|
- ...
|
|
|
- Frame 17 → cam20
|
|
|
- Frame 18 → cam00 (loops back)
|
|
|
- ...
|
|
|
- Frame 31 → cam14
|
|
|
|
|
|
## Data Processing
|
|
|
|
|
|
### Original Data
|
|
|
|
|
|
- **Source resolution**: 2704×2028 (4:3 aspect ratio)
|
|
|
- **Original format**: Multi-view MP4 videos
|
|
|
- **Camera model**: LLFF format with `poses_bounds.npy`
|
|
|
|
|
|
### Processing Pipeline
|
|
|
|
|
|
1. **Center Crop**: 2704×2028 → 2028×2028 (square)
|
|
|
2. **Resize**: 2028×2028 → 256×256 or 512×512
|
|
|
3. **Intrinsics Adjustment**: Focal length and principal point adjusted for crop and resize
|
|
|
4. **Extrinsics Extraction**: Camera poses extracted from LLFF format
|
|
|
5. **Format Conversion**: Converted to standard c2w/w2c matrices
|
|
|
|
|
|
### Frame Sampling Strategy
|
|
|
|
|
|
To capture the dynamic motion from multiple viewpoints, frames are sampled such that each frame shows the scene from a different camera angle in sequence. This creates a "synchronized" multi-view video where:
|
|
|
- The temporal progression shows the dynamic action
|
|
|
- Each frame provides a different spatial viewpoint
|
|
|
- Camera angles loop after exhausting all 18 cameras
|
|
|
|
|
|
## Camera Visualization
|
|
|
|
|
|
Each processed scene includes an interactive 3D camera visualization (`camera_visualization.html`):
|
|
|
|
|
|
- **View camera positions** and orientations in 3D space
|
|
|
- **Interactive**: Rotate, pan, and zoom to explore the camera rig
|
|
|
- **Camera frustums**: Visualize the viewing direction and field of view
|
|
|
- **Trajectory path**: See the sequence of frames and camera transitions
|
|
|
- **Powered by Plotly**: High-quality interactive graphics
|
|
|
|
|
|
Open the HTML file in any web browser to explore the camera setup.
|
|
|
|
|
|
## Usage Examples
|
|
|
|
|
|
### Loading Camera Poses (Python)
|
|
|
|
|
|
```python
|
|
|
import json
|
|
|
import numpy as np
|
|
|
|
|
|
# Load from JSON
|
|
|
with open('coffee_martini_processed/256/transforms.json', 'r') as f:
|
|
|
data = json.load(f)
|
|
|
|
|
|
# Access first frame
|
|
|
frame0 = data['frames'][0]['front']
|
|
|
print(f"Camera intrinsics: fx={frame0['fx']}, fy={frame0['fy']}")
|
|
|
print(f"Camera position: {frame0['blender_camera_location']}")
|
|
|
print(f"Image path: {frame0['file_path']}")
|
|
|
|
|
|
# Load from NumPy
|
|
|
poses = np.load('coffee_martini_processed/256/transforms.npz')
|
|
|
intrinsics = poses['intrinsics'] # (32, 3, 3)
|
|
|
c2w = poses['extrinsics_c2w'] # (32, 4, 4)
|
|
|
```
|
|
|
|
|
|
### Loading Images
|
|
|
|
|
|
```python
|
|
|
import cv2
|
|
|
import os
|
|
|
|
|
|
scene_dir = 'coffee_martini_processed/256'
|
|
|
img_dir = os.path.join(scene_dir, 'images')
|
|
|
|
|
|
# Load all frames
|
|
|
frames = []
|
|
|
for i in range(32):
|
|
|
img_path = os.path.join(img_dir, f'sample_{i:03d}_cam*.jpg')
|
|
|
# Find the actual file (camera number may vary)
|
|
|
import glob
|
|
|
img_file = glob.glob(img_path)[0]
|
|
|
img = cv2.imread(img_file)
|
|
|
frames.append(img)
|
|
|
|
|
|
print(f"Loaded {len(frames)} frames, shape: {frames[0].shape}")
|
|
|
```
|
|
|
|
|
|
### PyTorch Dataset Example
|
|
|
|
|
|
```python
|
|
|
import torch
|
|
|
from torch.utils.data import Dataset
|
|
|
from PIL import Image
|
|
|
import json
|
|
|
import numpy as np
|
|
|
|
|
|
class Neural3DVideoDataset(Dataset):
|
|
|
def __init__(self, scene_dir):
|
|
|
self.scene_dir = scene_dir
|
|
|
|
|
|
# Load transforms
|
|
|
with open(os.path.join(scene_dir, 'transforms.json'), 'r') as f:
|
|
|
self.data = json.load(f)
|
|
|
|
|
|
self.frames = self.data['frames']
|
|
|
|
|
|
def __len__(self):
|
|
|
return len(self.frames)
|
|
|
|
|
|
def __getitem__(self, idx):
|
|
|
frame_data = self.frames[idx]['front']
|
|
|
|
|
|
# Load image
|
|
|
img_path = os.path.join(self.scene_dir, frame_data['file_path'])
|
|
|
img = Image.open(img_path).convert('RGB')
|
|
|
img = torch.from_numpy(np.array(img)).float() / 255.0
|
|
|
|
|
|
# Get camera parameters
|
|
|
intrinsics = torch.tensor([
|
|
|
[frame_data['fx'], 0, frame_data['cx']],
|
|
|
[0, frame_data['fy'], frame_data['cy']],
|
|
|
[0, 0, 1]
|
|
|
], dtype=torch.float32)
|
|
|
|
|
|
c2w = torch.tensor(frame_data['c2w'], dtype=torch.float32)
|
|
|
|
|
|
return {
|
|
|
'image': img,
|
|
|
'intrinsics': intrinsics,
|
|
|
'c2w': c2w,
|
|
|
'timestamp': frame_data['timestamp']
|
|
|
}
|
|
|
|
|
|
# Usage
|
|
|
dataset = Neural3DVideoDataset('coffee_martini_processed/256')
|
|
|
sample = dataset[0]
|
|
|
print(f"Image shape: {sample['image'].shape}")
|
|
|
print(f"Camera position: {sample['c2w'][:, 3]}")
|
|
|
```
|
|
|
|
|
|
## Technical Details
|
|
|
|
|
|
### Camera Configuration
|
|
|
|
|
|
- **Number of cameras**: 18 per scene
|
|
|
- **Camera arrangement**: Surrounding the scene in a roughly circular pattern
|
|
|
- **Frame rate**: 30 FPS (original videos)
|
|
|
- **Camera model**: Pinhole camera with radial distortion (pre-undistorted)
|
|
|
|
|
|
### Coordinate System
|
|
|
|
|
|
- **World coordinates**: Right-handed coordinate system
|
|
|
- **Camera coordinates**:
|
|
|
- X-axis: Right
|
|
|
- Y-axis: Down
|
|
|
- Z-axis: Forward (viewing direction)
|
|
|
- **c2w matrix**: Transforms from camera space to world space
|
|
|
- **w2c matrix**: Transforms from world space to camera space
|
|
|
|
|
|
### Quality Settings
|
|
|
|
|
|
- **JPEG quality**: 95
|
|
|
- **Interpolation**: Bilinear (cv2.INTER_LINEAR)
|
|
|
- **Color space**: RGB (8-bit per channel)
|
|
|
|
|
|
## Citation
|
|
|
|
|
|
If you use this dataset in your research, please cite the original Neural 3D Video dataset:
|
|
|
|
|
|
```bibtex
|
|
|
@article{neural3dvideo2021,
|
|
|
title={Neural 3D Video Synthesis},
|
|
|
author={Author Names},
|
|
|
journal={Conference/Journal Name},
|
|
|
year={2021}
|
|
|
}
|
|
|
```
|
|
|
|
|
|
## Processing Scripts
|
|
|
|
|
|
The data was processed using custom scripts available in the parent directory:
|
|
|
- `create_sync_video_with_poses.py` - Single scene processing
|
|
|
- `batch_process_scenes.py` - Batch processing for all scenes
|
|
|
|
|
|
## License
|
|
|
|
|
|
This processed dataset inherits the CC-BY-4.0 license from the original Neural 3D Video dataset. Please respect the license terms when using this data.
|
|
|
|
|
|
## Contact
|
|
|
|
|
|
For questions or issues regarding this processed dataset, please contact the dataset maintainer. |