Xuweiyi's picture
upload
3797537
---
license: cc-by-4.0
---
# Neural 3D Video Dataset - Processed
This directory contains preprocessed multi-view video data from the Neural 3D Video dataset, converted into a format suitable for 4D reconstruction and novel view synthesis tasks.
## Dataset Overview
**Source Dataset**: Neural 3D Video
**License**: CC-BY-4.0
**Processed Scenes**: 5 dynamic cooking scenes captured from multiple camera angles
### Scenes
| Scene | Description | Cameras | Frames |
|-------|-------------|---------|--------|
| `coffee_martini` | Making a coffee martini cocktail | 18 | 32 |
| `cook_spinach` | Cooking spinach in a pan | 18 | 32 |
| `cut_roasted_beef` | Cutting roasted beef | 18 | 32 |
| `flame_salmon_1` | Flambé salmon preparation | 18 | 32 |
| `sear_steak` | Searing steak in a pan | 18 | 32 |
## Directory Structure
```
Neural-3D-Video-Dataset/
├── README.md (this file)
├── coffee_martini_processed/
│ ├── 256/ # 256×256 resolution
│ │ ├── images/ # 32 frame images
│ │ │ ├── sample_000_cam00.jpg
│ │ │ ├── sample_001_cam01.jpg
│ │ │ └── ...
│ │ ├── transforms.json # Camera poses (JSON format)
│ │ ├── transforms.npz # Camera poses (NumPy format)
│ │ └── camera_visualization.html # Interactive 3D camera viewer
│ └── 512/ # 512×512 resolution
│ ├── images/
│ ├── transforms.json
│ ├── transforms.npz
│ └── camera_visualization.html
├── cook_spinach_processed/
│ ├── 256/ ...
│ └── 512/ ...
├── cut_roasted_beef_processed/
│ ├── 256/ ...
│ └── 512/ ...
├── flame_salmon_1_processed/
│ ├── 256/ ...
│ └── 512/ ...
└── sear_steak_processed/
├── 256/ ...
└── 512/ ...
```
## Data Format
### Camera Poses (`transforms.json`)
The camera poses are stored in a JSON file with the following structure:
```json
{
"frames": [
{
"front": {
"timestamp": 0,
"file_path": "./images/sample_000_cam00.jpg",
"w": 256,
"h": 256,
"fx": 341.33,
"fy": 341.33,
"cx": 128.0,
"cy": 128.0,
"w2c": [[...], [...], [...], [...]], // 4×4 world-to-camera matrix
"c2w": [[...], [...], [...]], // 3×4 camera-to-world matrix
"blender_camera_location": [x, y, z] // Camera position in world coordinates
}
},
...
]
}
```
**Intrinsics** (camera internal parameters):
- **256×256**: `fx = fy = 341.33`, `cx = cy = 128.0`
- **512×512**: `fx = fy = 682.67`, `cx = cy = 256.0`
**Extrinsics** (camera external parameters):
- `w2c`: 4×4 world-to-camera transformation matrix
- `c2w`: 3×4 camera-to-world transformation matrix (rotation + translation)
- `blender_camera_location`: 3D camera position `[x, y, z]` in world coordinates
### NumPy Format (`transforms.npz`)
For convenience, camera parameters are also provided in NumPy format:
```python
import numpy as np
data = np.load('transforms.npz')
intrinsics = data['intrinsics'] # (32, 3, 3) - intrinsic matrices
extrinsics_w2c = data['extrinsics_w2c'] # (32, 4, 4) - world-to-camera
extrinsics_c2w = data['extrinsics_c2w'] # (32, 4, 4) - camera-to-world
camera_positions = data['camera_positions'] # (32, 3) - camera locations
```
### Frame Images
- **Format**: JPEG
- **Resolutions**: 256×256 and 512×512
- **Count**: 32 frames per scene
- **Naming**: `sample_{frame:03d}_cam{camera:02d}.jpg`
Each frame is extracted from a different camera view:
- Frame 0 → cam00
- Frame 1 → cam01
- ...
- Frame 17 → cam20
- Frame 18 → cam00 (loops back)
- ...
- Frame 31 → cam14
## Data Processing
### Original Data
- **Source resolution**: 2704×2028 (4:3 aspect ratio)
- **Original format**: Multi-view MP4 videos
- **Camera model**: LLFF format with `poses_bounds.npy`
### Processing Pipeline
1. **Center Crop**: 2704×2028 → 2028×2028 (square)
2. **Resize**: 2028×2028 → 256×256 or 512×512
3. **Intrinsics Adjustment**: Focal length and principal point adjusted for crop and resize
4. **Extrinsics Extraction**: Camera poses extracted from LLFF format
5. **Format Conversion**: Converted to standard c2w/w2c matrices
### Frame Sampling Strategy
To capture the dynamic motion from multiple viewpoints, frames are sampled such that each frame shows the scene from a different camera angle in sequence. This creates a "synchronized" multi-view video where:
- The temporal progression shows the dynamic action
- Each frame provides a different spatial viewpoint
- Camera angles loop after exhausting all 18 cameras
## Camera Visualization
Each processed scene includes an interactive 3D camera visualization (`camera_visualization.html`):
- **View camera positions** and orientations in 3D space
- **Interactive**: Rotate, pan, and zoom to explore the camera rig
- **Camera frustums**: Visualize the viewing direction and field of view
- **Trajectory path**: See the sequence of frames and camera transitions
- **Powered by Plotly**: High-quality interactive graphics
Open the HTML file in any web browser to explore the camera setup.
## Usage Examples
### Loading Camera Poses (Python)
```python
import json
import numpy as np
# Load from JSON
with open('coffee_martini_processed/256/transforms.json', 'r') as f:
data = json.load(f)
# Access first frame
frame0 = data['frames'][0]['front']
print(f"Camera intrinsics: fx={frame0['fx']}, fy={frame0['fy']}")
print(f"Camera position: {frame0['blender_camera_location']}")
print(f"Image path: {frame0['file_path']}")
# Load from NumPy
poses = np.load('coffee_martini_processed/256/transforms.npz')
intrinsics = poses['intrinsics'] # (32, 3, 3)
c2w = poses['extrinsics_c2w'] # (32, 4, 4)
```
### Loading Images
```python
import cv2
import os
scene_dir = 'coffee_martini_processed/256'
img_dir = os.path.join(scene_dir, 'images')
# Load all frames
frames = []
for i in range(32):
img_path = os.path.join(img_dir, f'sample_{i:03d}_cam*.jpg')
# Find the actual file (camera number may vary)
import glob
img_file = glob.glob(img_path)[0]
img = cv2.imread(img_file)
frames.append(img)
print(f"Loaded {len(frames)} frames, shape: {frames[0].shape}")
```
### PyTorch Dataset Example
```python
import torch
from torch.utils.data import Dataset
from PIL import Image
import json
import numpy as np
class Neural3DVideoDataset(Dataset):
def __init__(self, scene_dir):
self.scene_dir = scene_dir
# Load transforms
with open(os.path.join(scene_dir, 'transforms.json'), 'r') as f:
self.data = json.load(f)
self.frames = self.data['frames']
def __len__(self):
return len(self.frames)
def __getitem__(self, idx):
frame_data = self.frames[idx]['front']
# Load image
img_path = os.path.join(self.scene_dir, frame_data['file_path'])
img = Image.open(img_path).convert('RGB')
img = torch.from_numpy(np.array(img)).float() / 255.0
# Get camera parameters
intrinsics = torch.tensor([
[frame_data['fx'], 0, frame_data['cx']],
[0, frame_data['fy'], frame_data['cy']],
[0, 0, 1]
], dtype=torch.float32)
c2w = torch.tensor(frame_data['c2w'], dtype=torch.float32)
return {
'image': img,
'intrinsics': intrinsics,
'c2w': c2w,
'timestamp': frame_data['timestamp']
}
# Usage
dataset = Neural3DVideoDataset('coffee_martini_processed/256')
sample = dataset[0]
print(f"Image shape: {sample['image'].shape}")
print(f"Camera position: {sample['c2w'][:, 3]}")
```
## Technical Details
### Camera Configuration
- **Number of cameras**: 18 per scene
- **Camera arrangement**: Surrounding the scene in a roughly circular pattern
- **Frame rate**: 30 FPS (original videos)
- **Camera model**: Pinhole camera with radial distortion (pre-undistorted)
### Coordinate System
- **World coordinates**: Right-handed coordinate system
- **Camera coordinates**:
- X-axis: Right
- Y-axis: Down
- Z-axis: Forward (viewing direction)
- **c2w matrix**: Transforms from camera space to world space
- **w2c matrix**: Transforms from world space to camera space
### Quality Settings
- **JPEG quality**: 95
- **Interpolation**: Bilinear (cv2.INTER_LINEAR)
- **Color space**: RGB (8-bit per channel)
## Citation
If you use this dataset in your research, please cite the original Neural 3D Video dataset:
```bibtex
@article{neural3dvideo2021,
title={Neural 3D Video Synthesis},
author={Author Names},
journal={Conference/Journal Name},
year={2021}
}
```
## Processing Scripts
The data was processed using custom scripts available in the parent directory:
- `create_sync_video_with_poses.py` - Single scene processing
- `batch_process_scenes.py` - Batch processing for all scenes
## License
This processed dataset inherits the CC-BY-4.0 license from the original Neural 3D Video dataset. Please respect the license terms when using this data.
## Contact
For questions or issues regarding this processed dataset, please contact the dataset maintainer.