Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/__init__.py +0 -0
- external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/helpers.py +141 -0
- external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/imagebind_model.py +517 -0
- external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/multimodal_preprocessors.py +687 -0
- external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/transformer.py +284 -0
- external/Grounded-Segment-Anything/playground/PaintByExample/README.md +70 -0
- external/Grounded-Segment-Anything/playground/PaintByExample/paint_by_example.py +53 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/models/__init__.py +0 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/models/data_processor.py +211 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/models/mean_vfe.py +25 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/models/spconv_backbone_voxelnext.py +292 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/models/voxelnext_head.py +166 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/utils/centernet_utils.py +116 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/utils/config.py +34 -0
- external/Grounded-Segment-Anything/voxelnext_3d_box/utils/image_projection.py +103 -0
- external/WildCamera/WildCamera/benchmark/benchmark_calibration.py +98 -0
- external/WildCamera/WildCamera/benchmark/benchmark_crop.py +63 -0
- external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_megadepth.py +168 -0
- external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_scannet.py +165 -0
- external/WildCamera/WildCamera/datasets/GSV.py +73 -0
- external/WildCamera/WildCamera/datasets/GenericDataset.py +171 -0
- external/WildCamera/WildCamera/datasets/IncdDataset.py +110 -0
- external/WildCamera/WildCamera/datasets/MegaDepth.py +76 -0
- external/WildCamera/WildCamera/evaluation/evaluate_crop.py +157 -0
- external/WildCamera/WildCamera/evaluation/evaluate_fov.py +142 -0
- external/WildCamera/WildCamera/evaluation/evaluate_intrinsic.py +144 -0
- external/WildCamera/WildCamera/evaluation/evaluate_pose.py +209 -0
- external/WildCamera/WildCamera/newcrfs/__init__.py +0 -0
- external/WildCamera/WildCamera/newcrfs/__pycache__/swin_transformer.cpython-310.pyc +0 -0
- external/WildCamera/WildCamera/newcrfs/__pycache__/uper_crf_head.cpython-310.pyc +0 -0
- external/WildCamera/WildCamera/newcrfs/newcrf_incidencefield.py +196 -0
- external/WildCamera/WildCamera/newcrfs/newcrf_layers.py +433 -0
- external/WildCamera/WildCamera/newcrfs/newcrf_utils.py +264 -0
- external/WildCamera/WildCamera/newcrfs/swin_transformer.py +619 -0
- external/WildCamera/WildCamera/newcrfs/uper_crf_head.py +350 -0
- external/WildCamera/WildCamera/train/train_calibrator.py +297 -0
- external/WildCamera/__pycache__/hubconf.cpython-310.pyc +0 -0
- external/WildCamera/asset/download_demo_images.sh +8 -0
- external/WildCamera/asset/download_wildcamera_checkpoint.sh +13 -0
- external/WildCamera/asset/download_wildcamera_dataset.sh +0 -0
- external/WildCamera/demo/demo_dollyzoom.py +133 -0
- external/WildCamera/demo/demo_inference.py +30 -0
- external/WildCamera/demo/demo_restoration.py +33 -0
- external/WildCamera/splits/arkitscenes_test.txt +800 -0
- external/WildCamera/splits/arkitscenes_train.txt +0 -0
- external/WildCamera/splits/arkitscenes_val.txt +800 -0
- external/WildCamera/splits/biwirgbdid_test.txt +800 -0
- external/WildCamera/splits/cad120_test.txt +800 -0
- external/WildCamera/splits/cityscapes_test.txt +800 -0
- external/WildCamera/splits/cityscapes_train.txt +0 -0
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/__init__.py
ADDED
|
File without changes
|
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/helpers.py
ADDED
|
@@ -0,0 +1,141 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env python3
|
| 2 |
+
# Portions Copyright (c) Meta Platforms, Inc. and affiliates.
|
| 3 |
+
# All rights reserved.
|
| 4 |
+
|
| 5 |
+
# This source code is licensed under the license found in the
|
| 6 |
+
# LICENSE file in the root directory of this source tree.
|
| 7 |
+
|
| 8 |
+
import math
|
| 9 |
+
|
| 10 |
+
import einops
|
| 11 |
+
import numpy as np
|
| 12 |
+
import torch
|
| 13 |
+
|
| 14 |
+
import torch.nn as nn
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
class Normalize(nn.Module):
|
| 18 |
+
def __init__(self, dim: int) -> None:
|
| 19 |
+
super().__init__()
|
| 20 |
+
self.dim = dim
|
| 21 |
+
|
| 22 |
+
def forward(self, x):
|
| 23 |
+
return torch.nn.functional.normalize(x, dim=self.dim, p=2)
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
class LearnableLogitScaling(nn.Module):
|
| 27 |
+
def __init__(
|
| 28 |
+
self,
|
| 29 |
+
logit_scale_init: float = 1 / 0.07,
|
| 30 |
+
learnable: bool = True,
|
| 31 |
+
max_logit_scale: float = 100,
|
| 32 |
+
) -> None:
|
| 33 |
+
super().__init__()
|
| 34 |
+
self.max_logit_scale = max_logit_scale
|
| 35 |
+
self.logit_scale_init = logit_scale_init
|
| 36 |
+
self.learnable = learnable
|
| 37 |
+
log_logit_scale = torch.ones([]) * np.log(self.logit_scale_init)
|
| 38 |
+
if learnable:
|
| 39 |
+
self.log_logit_scale = nn.Parameter(log_logit_scale)
|
| 40 |
+
else:
|
| 41 |
+
self.register_buffer("log_logit_scale", log_logit_scale)
|
| 42 |
+
|
| 43 |
+
def forward(self, x):
|
| 44 |
+
return torch.clip(self.log_logit_scale.exp(), max=self.max_logit_scale) * x
|
| 45 |
+
|
| 46 |
+
def extra_repr(self):
|
| 47 |
+
st = f"logit_scale_init={self.logit_scale_init},learnable={self.learnable}, max_logit_scale={self.max_logit_scale}"
|
| 48 |
+
return st
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
class EinOpsRearrange(nn.Module):
|
| 52 |
+
def __init__(self, rearrange_expr: str, **kwargs) -> None:
|
| 53 |
+
super().__init__()
|
| 54 |
+
self.rearrange_expr = rearrange_expr
|
| 55 |
+
self.kwargs = kwargs
|
| 56 |
+
|
| 57 |
+
def forward(self, x):
|
| 58 |
+
assert isinstance(x, torch.Tensor)
|
| 59 |
+
return einops.rearrange(x, self.rearrange_expr, **self.kwargs)
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
class VerboseNNModule(nn.Module):
|
| 63 |
+
"""
|
| 64 |
+
Wrapper around nn.Module that prints registered buffers and parameter names.
|
| 65 |
+
"""
|
| 66 |
+
|
| 67 |
+
@staticmethod
|
| 68 |
+
def get_readable_tensor_repr(name: str, tensor: torch.Tensor) -> str:
|
| 69 |
+
st = (
|
| 70 |
+
"("
|
| 71 |
+
+ name
|
| 72 |
+
+ "): "
|
| 73 |
+
+ "tensor("
|
| 74 |
+
+ str(tuple(tensor[1].shape))
|
| 75 |
+
+ ", requires_grad="
|
| 76 |
+
+ str(tensor[1].requires_grad)
|
| 77 |
+
+ ")\n"
|
| 78 |
+
)
|
| 79 |
+
return st
|
| 80 |
+
|
| 81 |
+
def extra_repr(self) -> str:
|
| 82 |
+
named_modules = set()
|
| 83 |
+
for p in self.named_modules():
|
| 84 |
+
named_modules.update([p[0]])
|
| 85 |
+
named_modules = list(named_modules)
|
| 86 |
+
|
| 87 |
+
string_repr = ""
|
| 88 |
+
for p in self.named_parameters():
|
| 89 |
+
name = p[0].split(".")[0]
|
| 90 |
+
if name not in named_modules:
|
| 91 |
+
string_repr += self.get_readable_tensor_repr(name, p)
|
| 92 |
+
|
| 93 |
+
for p in self.named_buffers():
|
| 94 |
+
name = p[0].split(".")[0]
|
| 95 |
+
string_repr += self.get_readable_tensor_repr(name, p)
|
| 96 |
+
|
| 97 |
+
return string_repr
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
def cast_if_src_dtype(
|
| 101 |
+
tensor: torch.Tensor, src_dtype: torch.dtype, tgt_dtype: torch.dtype
|
| 102 |
+
):
|
| 103 |
+
updated = False
|
| 104 |
+
if tensor.dtype == src_dtype:
|
| 105 |
+
tensor = tensor.to(dtype=tgt_dtype)
|
| 106 |
+
updated = True
|
| 107 |
+
return tensor, updated
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
class QuickGELU(nn.Module):
|
| 111 |
+
# From https://github.com/openai/CLIP/blob/d50d76daa670286dd6cacf3bcd80b5e4823fc8e1/clip/model.py#L166
|
| 112 |
+
def forward(self, x: torch.Tensor):
|
| 113 |
+
return x * torch.sigmoid(1.702 * x)
|
| 114 |
+
|
| 115 |
+
|
| 116 |
+
class SelectElement(nn.Module):
|
| 117 |
+
def __init__(self, index) -> None:
|
| 118 |
+
super().__init__()
|
| 119 |
+
self.index = index
|
| 120 |
+
|
| 121 |
+
def forward(self, x):
|
| 122 |
+
assert x.ndim >= 3
|
| 123 |
+
return x[:, self.index, ...]
|
| 124 |
+
|
| 125 |
+
|
| 126 |
+
class SelectEOSAndProject(nn.Module):
|
| 127 |
+
"""
|
| 128 |
+
Text Pooling used in OpenCLIP
|
| 129 |
+
"""
|
| 130 |
+
|
| 131 |
+
def __init__(self, proj: nn.Module) -> None:
|
| 132 |
+
super().__init__()
|
| 133 |
+
self.proj = proj
|
| 134 |
+
|
| 135 |
+
def forward(self, x, seq_len):
|
| 136 |
+
assert x.ndim == 3
|
| 137 |
+
# x is of shape B x L x D
|
| 138 |
+
# take features from the eot embedding (eot_token is the highest number in each sequence)
|
| 139 |
+
x = x[torch.arange(x.shape[0]), seq_len]
|
| 140 |
+
x = self.proj(x)
|
| 141 |
+
return x
|
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/imagebind_model.py
ADDED
|
@@ -0,0 +1,517 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env python3
|
| 2 |
+
# Portions Copyright (c) Meta Platforms, Inc. and affiliates.
|
| 3 |
+
# All rights reserved.
|
| 4 |
+
|
| 5 |
+
# This source code is licensed under the license found in the
|
| 6 |
+
# LICENSE file in the root directory of this source tree.
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
import os
|
| 10 |
+
import urllib
|
| 11 |
+
from functools import partial
|
| 12 |
+
from types import SimpleNamespace
|
| 13 |
+
|
| 14 |
+
import torch
|
| 15 |
+
import torch.nn as nn
|
| 16 |
+
|
| 17 |
+
from models.helpers import (
|
| 18 |
+
EinOpsRearrange,
|
| 19 |
+
LearnableLogitScaling,
|
| 20 |
+
Normalize,
|
| 21 |
+
SelectElement,
|
| 22 |
+
SelectEOSAndProject,
|
| 23 |
+
)
|
| 24 |
+
from models.multimodal_preprocessors import (
|
| 25 |
+
AudioPreprocessor,
|
| 26 |
+
IMUPreprocessor,
|
| 27 |
+
PadIm2Video,
|
| 28 |
+
PatchEmbedGeneric,
|
| 29 |
+
RGBDTPreprocessor,
|
| 30 |
+
SpatioTemporalPosEmbeddingHelper,
|
| 31 |
+
TextPreprocessor,
|
| 32 |
+
ThermalPreprocessor,
|
| 33 |
+
)
|
| 34 |
+
|
| 35 |
+
from models.transformer import MultiheadAttention, SimpleTransformer
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
ModalityType = SimpleNamespace(
|
| 39 |
+
VISION="vision",
|
| 40 |
+
TEXT="text",
|
| 41 |
+
AUDIO="audio",
|
| 42 |
+
THERMAL="thermal",
|
| 43 |
+
DEPTH="depth",
|
| 44 |
+
IMU="imu",
|
| 45 |
+
)
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
class ImageBindModel(nn.Module):
|
| 49 |
+
def __init__(
|
| 50 |
+
self,
|
| 51 |
+
video_frames=2,
|
| 52 |
+
kernel_size=(2, 14, 14),
|
| 53 |
+
audio_kernel_size=16,
|
| 54 |
+
audio_stride=10,
|
| 55 |
+
out_embed_dim=768,
|
| 56 |
+
vision_embed_dim=1024,
|
| 57 |
+
vision_num_blocks=24,
|
| 58 |
+
vision_num_heads=16,
|
| 59 |
+
audio_embed_dim=768,
|
| 60 |
+
audio_num_blocks=12,
|
| 61 |
+
audio_num_heads=12,
|
| 62 |
+
audio_num_mel_bins=128,
|
| 63 |
+
audio_target_len=204,
|
| 64 |
+
audio_drop_path=0.1,
|
| 65 |
+
text_embed_dim=768,
|
| 66 |
+
text_num_blocks=12,
|
| 67 |
+
text_num_heads=12,
|
| 68 |
+
depth_embed_dim=384,
|
| 69 |
+
depth_kernel_size=16,
|
| 70 |
+
depth_num_blocks=12,
|
| 71 |
+
depth_num_heads=8,
|
| 72 |
+
depth_drop_path=0.0,
|
| 73 |
+
thermal_embed_dim=768,
|
| 74 |
+
thermal_kernel_size=16,
|
| 75 |
+
thermal_num_blocks=12,
|
| 76 |
+
thermal_num_heads=12,
|
| 77 |
+
thermal_drop_path=0.0,
|
| 78 |
+
imu_embed_dim=512,
|
| 79 |
+
imu_kernel_size=8,
|
| 80 |
+
imu_num_blocks=6,
|
| 81 |
+
imu_num_heads=8,
|
| 82 |
+
imu_drop_path=0.7,
|
| 83 |
+
):
|
| 84 |
+
super().__init__()
|
| 85 |
+
|
| 86 |
+
self.modality_preprocessors = self._create_modality_preprocessors(
|
| 87 |
+
video_frames,
|
| 88 |
+
vision_embed_dim,
|
| 89 |
+
kernel_size,
|
| 90 |
+
text_embed_dim,
|
| 91 |
+
audio_embed_dim,
|
| 92 |
+
audio_kernel_size,
|
| 93 |
+
audio_stride,
|
| 94 |
+
audio_num_mel_bins,
|
| 95 |
+
audio_target_len,
|
| 96 |
+
depth_embed_dim,
|
| 97 |
+
depth_kernel_size,
|
| 98 |
+
thermal_embed_dim,
|
| 99 |
+
thermal_kernel_size,
|
| 100 |
+
imu_embed_dim,
|
| 101 |
+
)
|
| 102 |
+
|
| 103 |
+
self.modality_trunks = self._create_modality_trunks(
|
| 104 |
+
vision_embed_dim,
|
| 105 |
+
vision_num_blocks,
|
| 106 |
+
vision_num_heads,
|
| 107 |
+
text_embed_dim,
|
| 108 |
+
text_num_blocks,
|
| 109 |
+
text_num_heads,
|
| 110 |
+
audio_embed_dim,
|
| 111 |
+
audio_num_blocks,
|
| 112 |
+
audio_num_heads,
|
| 113 |
+
audio_drop_path,
|
| 114 |
+
depth_embed_dim,
|
| 115 |
+
depth_num_blocks,
|
| 116 |
+
depth_num_heads,
|
| 117 |
+
depth_drop_path,
|
| 118 |
+
thermal_embed_dim,
|
| 119 |
+
thermal_num_blocks,
|
| 120 |
+
thermal_num_heads,
|
| 121 |
+
thermal_drop_path,
|
| 122 |
+
imu_embed_dim,
|
| 123 |
+
imu_num_blocks,
|
| 124 |
+
imu_num_heads,
|
| 125 |
+
imu_drop_path,
|
| 126 |
+
)
|
| 127 |
+
|
| 128 |
+
self.modality_heads = self._create_modality_heads(
|
| 129 |
+
out_embed_dim,
|
| 130 |
+
vision_embed_dim,
|
| 131 |
+
text_embed_dim,
|
| 132 |
+
audio_embed_dim,
|
| 133 |
+
depth_embed_dim,
|
| 134 |
+
thermal_embed_dim,
|
| 135 |
+
imu_embed_dim,
|
| 136 |
+
)
|
| 137 |
+
|
| 138 |
+
self.modality_postprocessors = self._create_modality_postprocessors(
|
| 139 |
+
out_embed_dim
|
| 140 |
+
)
|
| 141 |
+
|
| 142 |
+
def _create_modality_preprocessors(
|
| 143 |
+
self,
|
| 144 |
+
video_frames=2,
|
| 145 |
+
vision_embed_dim=1024,
|
| 146 |
+
kernel_size=(2, 14, 14),
|
| 147 |
+
text_embed_dim=768,
|
| 148 |
+
audio_embed_dim=768,
|
| 149 |
+
audio_kernel_size=16,
|
| 150 |
+
audio_stride=10,
|
| 151 |
+
audio_num_mel_bins=128,
|
| 152 |
+
audio_target_len=204,
|
| 153 |
+
depth_embed_dim=768,
|
| 154 |
+
depth_kernel_size=16,
|
| 155 |
+
thermal_embed_dim=768,
|
| 156 |
+
thermal_kernel_size=16,
|
| 157 |
+
imu_embed_dim=512,
|
| 158 |
+
):
|
| 159 |
+
rgbt_stem = PatchEmbedGeneric(
|
| 160 |
+
proj_stem=[
|
| 161 |
+
PadIm2Video(pad_type="repeat", ntimes=2),
|
| 162 |
+
nn.Conv3d(
|
| 163 |
+
in_channels=3,
|
| 164 |
+
kernel_size=kernel_size,
|
| 165 |
+
out_channels=vision_embed_dim,
|
| 166 |
+
stride=kernel_size,
|
| 167 |
+
bias=False,
|
| 168 |
+
),
|
| 169 |
+
]
|
| 170 |
+
)
|
| 171 |
+
rgbt_preprocessor = RGBDTPreprocessor(
|
| 172 |
+
img_size=[3, video_frames, 224, 224],
|
| 173 |
+
num_cls_tokens=1,
|
| 174 |
+
pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
|
| 175 |
+
rgbt_stem=rgbt_stem,
|
| 176 |
+
depth_stem=None,
|
| 177 |
+
)
|
| 178 |
+
|
| 179 |
+
text_preprocessor = TextPreprocessor(
|
| 180 |
+
context_length=77,
|
| 181 |
+
vocab_size=49408,
|
| 182 |
+
embed_dim=text_embed_dim,
|
| 183 |
+
causal_masking=True,
|
| 184 |
+
)
|
| 185 |
+
|
| 186 |
+
audio_stem = PatchEmbedGeneric(
|
| 187 |
+
proj_stem=[
|
| 188 |
+
nn.Conv2d(
|
| 189 |
+
in_channels=1,
|
| 190 |
+
kernel_size=audio_kernel_size,
|
| 191 |
+
stride=audio_stride,
|
| 192 |
+
out_channels=audio_embed_dim,
|
| 193 |
+
bias=False,
|
| 194 |
+
),
|
| 195 |
+
],
|
| 196 |
+
norm_layer=nn.LayerNorm(normalized_shape=audio_embed_dim),
|
| 197 |
+
)
|
| 198 |
+
audio_preprocessor = AudioPreprocessor(
|
| 199 |
+
img_size=[1, audio_num_mel_bins, audio_target_len],
|
| 200 |
+
num_cls_tokens=1,
|
| 201 |
+
pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
|
| 202 |
+
audio_stem=audio_stem,
|
| 203 |
+
)
|
| 204 |
+
|
| 205 |
+
depth_stem = PatchEmbedGeneric(
|
| 206 |
+
[
|
| 207 |
+
nn.Conv2d(
|
| 208 |
+
kernel_size=depth_kernel_size,
|
| 209 |
+
in_channels=1,
|
| 210 |
+
out_channels=depth_embed_dim,
|
| 211 |
+
stride=depth_kernel_size,
|
| 212 |
+
bias=False,
|
| 213 |
+
),
|
| 214 |
+
],
|
| 215 |
+
norm_layer=nn.LayerNorm(normalized_shape=depth_embed_dim),
|
| 216 |
+
)
|
| 217 |
+
|
| 218 |
+
depth_preprocessor = RGBDTPreprocessor(
|
| 219 |
+
img_size=[1, 224, 224],
|
| 220 |
+
num_cls_tokens=1,
|
| 221 |
+
pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
|
| 222 |
+
rgbt_stem=None,
|
| 223 |
+
depth_stem=depth_stem,
|
| 224 |
+
)
|
| 225 |
+
|
| 226 |
+
thermal_stem = PatchEmbedGeneric(
|
| 227 |
+
[
|
| 228 |
+
nn.Conv2d(
|
| 229 |
+
kernel_size=thermal_kernel_size,
|
| 230 |
+
in_channels=1,
|
| 231 |
+
out_channels=thermal_embed_dim,
|
| 232 |
+
stride=thermal_kernel_size,
|
| 233 |
+
bias=False,
|
| 234 |
+
),
|
| 235 |
+
],
|
| 236 |
+
norm_layer=nn.LayerNorm(normalized_shape=thermal_embed_dim),
|
| 237 |
+
)
|
| 238 |
+
thermal_preprocessor = ThermalPreprocessor(
|
| 239 |
+
img_size=[1, 224, 224],
|
| 240 |
+
num_cls_tokens=1,
|
| 241 |
+
pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
|
| 242 |
+
thermal_stem=thermal_stem,
|
| 243 |
+
)
|
| 244 |
+
|
| 245 |
+
imu_stem = PatchEmbedGeneric(
|
| 246 |
+
[
|
| 247 |
+
nn.Linear(
|
| 248 |
+
in_features=48,
|
| 249 |
+
out_features=imu_embed_dim,
|
| 250 |
+
bias=False,
|
| 251 |
+
),
|
| 252 |
+
],
|
| 253 |
+
norm_layer=nn.LayerNorm(normalized_shape=imu_embed_dim),
|
| 254 |
+
)
|
| 255 |
+
|
| 256 |
+
imu_preprocessor = IMUPreprocessor(
|
| 257 |
+
img_size=[6, 2000],
|
| 258 |
+
num_cls_tokens=1,
|
| 259 |
+
kernel_size=8,
|
| 260 |
+
embed_dim=imu_embed_dim,
|
| 261 |
+
pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
|
| 262 |
+
imu_stem=imu_stem,
|
| 263 |
+
)
|
| 264 |
+
|
| 265 |
+
modality_preprocessors = {
|
| 266 |
+
ModalityType.VISION: rgbt_preprocessor,
|
| 267 |
+
ModalityType.TEXT: text_preprocessor,
|
| 268 |
+
ModalityType.AUDIO: audio_preprocessor,
|
| 269 |
+
ModalityType.DEPTH: depth_preprocessor,
|
| 270 |
+
ModalityType.THERMAL: thermal_preprocessor,
|
| 271 |
+
ModalityType.IMU: imu_preprocessor,
|
| 272 |
+
}
|
| 273 |
+
|
| 274 |
+
return nn.ModuleDict(modality_preprocessors)
|
| 275 |
+
|
| 276 |
+
def _create_modality_trunks(
|
| 277 |
+
self,
|
| 278 |
+
vision_embed_dim=1024,
|
| 279 |
+
vision_num_blocks=24,
|
| 280 |
+
vision_num_heads=16,
|
| 281 |
+
text_embed_dim=768,
|
| 282 |
+
text_num_blocks=12,
|
| 283 |
+
text_num_heads=12,
|
| 284 |
+
audio_embed_dim=768,
|
| 285 |
+
audio_num_blocks=12,
|
| 286 |
+
audio_num_heads=12,
|
| 287 |
+
audio_drop_path=0.0,
|
| 288 |
+
depth_embed_dim=768,
|
| 289 |
+
depth_num_blocks=12,
|
| 290 |
+
depth_num_heads=12,
|
| 291 |
+
depth_drop_path=0.0,
|
| 292 |
+
thermal_embed_dim=768,
|
| 293 |
+
thermal_num_blocks=12,
|
| 294 |
+
thermal_num_heads=12,
|
| 295 |
+
thermal_drop_path=0.0,
|
| 296 |
+
imu_embed_dim=512,
|
| 297 |
+
imu_num_blocks=6,
|
| 298 |
+
imu_num_heads=8,
|
| 299 |
+
imu_drop_path=0.7,
|
| 300 |
+
):
|
| 301 |
+
def instantiate_trunk(
|
| 302 |
+
embed_dim, num_blocks, num_heads, pre_transformer_ln, add_bias_kv, drop_path
|
| 303 |
+
):
|
| 304 |
+
return SimpleTransformer(
|
| 305 |
+
embed_dim=embed_dim,
|
| 306 |
+
num_blocks=num_blocks,
|
| 307 |
+
ffn_dropout_rate=0.0,
|
| 308 |
+
drop_path_rate=drop_path,
|
| 309 |
+
attn_target=partial(
|
| 310 |
+
MultiheadAttention,
|
| 311 |
+
embed_dim=embed_dim,
|
| 312 |
+
num_heads=num_heads,
|
| 313 |
+
bias=True,
|
| 314 |
+
add_bias_kv=add_bias_kv,
|
| 315 |
+
),
|
| 316 |
+
pre_transformer_layer=nn.Sequential(
|
| 317 |
+
nn.LayerNorm(embed_dim, eps=1e-6)
|
| 318 |
+
if pre_transformer_ln
|
| 319 |
+
else nn.Identity(),
|
| 320 |
+
EinOpsRearrange("b l d -> l b d"),
|
| 321 |
+
),
|
| 322 |
+
post_transformer_layer=EinOpsRearrange("l b d -> b l d"),
|
| 323 |
+
)
|
| 324 |
+
|
| 325 |
+
modality_trunks = {}
|
| 326 |
+
modality_trunks[ModalityType.VISION] = instantiate_trunk(
|
| 327 |
+
vision_embed_dim,
|
| 328 |
+
vision_num_blocks,
|
| 329 |
+
vision_num_heads,
|
| 330 |
+
pre_transformer_ln=True,
|
| 331 |
+
add_bias_kv=False,
|
| 332 |
+
drop_path=0.0,
|
| 333 |
+
)
|
| 334 |
+
modality_trunks[ModalityType.TEXT] = instantiate_trunk(
|
| 335 |
+
text_embed_dim,
|
| 336 |
+
text_num_blocks,
|
| 337 |
+
text_num_heads,
|
| 338 |
+
pre_transformer_ln=False,
|
| 339 |
+
add_bias_kv=False,
|
| 340 |
+
drop_path=0.0,
|
| 341 |
+
)
|
| 342 |
+
modality_trunks[ModalityType.AUDIO] = instantiate_trunk(
|
| 343 |
+
audio_embed_dim,
|
| 344 |
+
audio_num_blocks,
|
| 345 |
+
audio_num_heads,
|
| 346 |
+
pre_transformer_ln=False,
|
| 347 |
+
add_bias_kv=True,
|
| 348 |
+
drop_path=audio_drop_path,
|
| 349 |
+
)
|
| 350 |
+
modality_trunks[ModalityType.DEPTH] = instantiate_trunk(
|
| 351 |
+
depth_embed_dim,
|
| 352 |
+
depth_num_blocks,
|
| 353 |
+
depth_num_heads,
|
| 354 |
+
pre_transformer_ln=False,
|
| 355 |
+
add_bias_kv=True,
|
| 356 |
+
drop_path=depth_drop_path,
|
| 357 |
+
)
|
| 358 |
+
modality_trunks[ModalityType.THERMAL] = instantiate_trunk(
|
| 359 |
+
thermal_embed_dim,
|
| 360 |
+
thermal_num_blocks,
|
| 361 |
+
thermal_num_heads,
|
| 362 |
+
pre_transformer_ln=False,
|
| 363 |
+
add_bias_kv=True,
|
| 364 |
+
drop_path=thermal_drop_path,
|
| 365 |
+
)
|
| 366 |
+
modality_trunks[ModalityType.IMU] = instantiate_trunk(
|
| 367 |
+
imu_embed_dim,
|
| 368 |
+
imu_num_blocks,
|
| 369 |
+
imu_num_heads,
|
| 370 |
+
pre_transformer_ln=False,
|
| 371 |
+
add_bias_kv=True,
|
| 372 |
+
drop_path=imu_drop_path,
|
| 373 |
+
)
|
| 374 |
+
|
| 375 |
+
return nn.ModuleDict(modality_trunks)
|
| 376 |
+
|
| 377 |
+
def _create_modality_heads(
|
| 378 |
+
self,
|
| 379 |
+
out_embed_dim,
|
| 380 |
+
vision_embed_dim,
|
| 381 |
+
text_embed_dim,
|
| 382 |
+
audio_embed_dim,
|
| 383 |
+
depth_embed_dim,
|
| 384 |
+
thermal_embed_dim,
|
| 385 |
+
imu_embed_dim,
|
| 386 |
+
):
|
| 387 |
+
modality_heads = {}
|
| 388 |
+
|
| 389 |
+
modality_heads[ModalityType.VISION] = nn.Sequential(
|
| 390 |
+
nn.LayerNorm(normalized_shape=vision_embed_dim, eps=1e-6),
|
| 391 |
+
SelectElement(index=0),
|
| 392 |
+
nn.Linear(vision_embed_dim, out_embed_dim, bias=False),
|
| 393 |
+
)
|
| 394 |
+
|
| 395 |
+
modality_heads[ModalityType.TEXT] = SelectEOSAndProject(
|
| 396 |
+
proj=nn.Sequential(
|
| 397 |
+
nn.LayerNorm(normalized_shape=text_embed_dim, eps=1e-6),
|
| 398 |
+
nn.Linear(text_embed_dim, out_embed_dim, bias=False),
|
| 399 |
+
)
|
| 400 |
+
)
|
| 401 |
+
|
| 402 |
+
modality_heads[ModalityType.AUDIO] = nn.Sequential(
|
| 403 |
+
nn.LayerNorm(normalized_shape=audio_embed_dim, eps=1e-6),
|
| 404 |
+
SelectElement(index=0),
|
| 405 |
+
nn.Linear(audio_embed_dim, out_embed_dim, bias=False),
|
| 406 |
+
)
|
| 407 |
+
|
| 408 |
+
modality_heads[ModalityType.DEPTH] = nn.Sequential(
|
| 409 |
+
nn.LayerNorm(normalized_shape=depth_embed_dim, eps=1e-6),
|
| 410 |
+
SelectElement(index=0),
|
| 411 |
+
nn.Linear(depth_embed_dim, out_embed_dim, bias=False),
|
| 412 |
+
)
|
| 413 |
+
|
| 414 |
+
modality_heads[ModalityType.THERMAL] = nn.Sequential(
|
| 415 |
+
nn.LayerNorm(normalized_shape=thermal_embed_dim, eps=1e-6),
|
| 416 |
+
SelectElement(index=0),
|
| 417 |
+
nn.Linear(thermal_embed_dim, out_embed_dim, bias=False),
|
| 418 |
+
)
|
| 419 |
+
|
| 420 |
+
modality_heads[ModalityType.IMU] = nn.Sequential(
|
| 421 |
+
nn.LayerNorm(normalized_shape=imu_embed_dim, eps=1e-6),
|
| 422 |
+
SelectElement(index=0),
|
| 423 |
+
nn.Dropout(p=0.5),
|
| 424 |
+
nn.Linear(imu_embed_dim, out_embed_dim, bias=False),
|
| 425 |
+
)
|
| 426 |
+
|
| 427 |
+
return nn.ModuleDict(modality_heads)
|
| 428 |
+
|
| 429 |
+
def _create_modality_postprocessors(self, out_embed_dim):
|
| 430 |
+
modality_postprocessors = {}
|
| 431 |
+
|
| 432 |
+
modality_postprocessors[ModalityType.VISION] = Normalize(dim=-1)
|
| 433 |
+
modality_postprocessors[ModalityType.TEXT] = nn.Sequential(
|
| 434 |
+
Normalize(dim=-1), LearnableLogitScaling(learnable=True)
|
| 435 |
+
)
|
| 436 |
+
modality_postprocessors[ModalityType.AUDIO] = nn.Sequential(
|
| 437 |
+
Normalize(dim=-1),
|
| 438 |
+
LearnableLogitScaling(logit_scale_init=20.0, learnable=False),
|
| 439 |
+
)
|
| 440 |
+
modality_postprocessors[ModalityType.DEPTH] = nn.Sequential(
|
| 441 |
+
Normalize(dim=-1),
|
| 442 |
+
LearnableLogitScaling(logit_scale_init=5.0, learnable=False),
|
| 443 |
+
)
|
| 444 |
+
modality_postprocessors[ModalityType.THERMAL] = nn.Sequential(
|
| 445 |
+
Normalize(dim=-1),
|
| 446 |
+
LearnableLogitScaling(logit_scale_init=10.0, learnable=False),
|
| 447 |
+
)
|
| 448 |
+
modality_postprocessors[ModalityType.IMU] = nn.Sequential(
|
| 449 |
+
Normalize(dim=-1),
|
| 450 |
+
LearnableLogitScaling(logit_scale_init=5.0, learnable=False),
|
| 451 |
+
)
|
| 452 |
+
|
| 453 |
+
return nn.ModuleDict(modality_postprocessors)
|
| 454 |
+
|
| 455 |
+
def forward(self, inputs):
|
| 456 |
+
outputs = {}
|
| 457 |
+
for modality_key, modality_value in inputs.items():
|
| 458 |
+
reduce_list = (
|
| 459 |
+
modality_value.ndim >= 5
|
| 460 |
+
) # Audio and Video inputs consist of multiple clips
|
| 461 |
+
if reduce_list:
|
| 462 |
+
B, S = modality_value.shape[:2]
|
| 463 |
+
modality_value = modality_value.reshape(
|
| 464 |
+
B * S, *modality_value.shape[2:]
|
| 465 |
+
)
|
| 466 |
+
|
| 467 |
+
if modality_value is not None:
|
| 468 |
+
modality_value = self.modality_preprocessors[modality_key](
|
| 469 |
+
**{modality_key: modality_value}
|
| 470 |
+
)
|
| 471 |
+
trunk_inputs = modality_value["trunk"]
|
| 472 |
+
head_inputs = modality_value["head"]
|
| 473 |
+
modality_value = self.modality_trunks[modality_key](**trunk_inputs)
|
| 474 |
+
modality_value = self.modality_heads[modality_key](
|
| 475 |
+
modality_value, **head_inputs
|
| 476 |
+
)
|
| 477 |
+
modality_value = self.modality_postprocessors[modality_key](
|
| 478 |
+
modality_value
|
| 479 |
+
)
|
| 480 |
+
|
| 481 |
+
if reduce_list:
|
| 482 |
+
modality_value = modality_value.reshape(B, S, -1)
|
| 483 |
+
modality_value = modality_value.mean(dim=1)
|
| 484 |
+
|
| 485 |
+
outputs[modality_key] = modality_value
|
| 486 |
+
|
| 487 |
+
return outputs
|
| 488 |
+
|
| 489 |
+
|
| 490 |
+
def imagebind_huge(pretrained=False):
|
| 491 |
+
model = ImageBindModel(
|
| 492 |
+
vision_embed_dim=1280,
|
| 493 |
+
vision_num_blocks=32,
|
| 494 |
+
vision_num_heads=16,
|
| 495 |
+
text_embed_dim=1024,
|
| 496 |
+
text_num_blocks=24,
|
| 497 |
+
text_num_heads=16,
|
| 498 |
+
out_embed_dim=1024,
|
| 499 |
+
audio_drop_path=0.1,
|
| 500 |
+
imu_drop_path=0.7,
|
| 501 |
+
)
|
| 502 |
+
|
| 503 |
+
if pretrained:
|
| 504 |
+
if not os.path.exists(".checkpoints/imagebind_huge.pth"):
|
| 505 |
+
print(
|
| 506 |
+
"Downloading imagebind weights to .checkpoints/imagebind_huge.pth ..."
|
| 507 |
+
)
|
| 508 |
+
os.makedirs(".checkpoints", exist_ok=True)
|
| 509 |
+
torch.hub.download_url_to_file(
|
| 510 |
+
"https://dl.fbaipublicfiles.com/imagebind/imagebind_huge.pth",
|
| 511 |
+
".checkpoints/imagebind_huge.pth",
|
| 512 |
+
progress=True,
|
| 513 |
+
)
|
| 514 |
+
|
| 515 |
+
model.load_state_dict(torch.load(".checkpoints/imagebind_huge.pth"))
|
| 516 |
+
|
| 517 |
+
return model
|
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/multimodal_preprocessors.py
ADDED
|
@@ -0,0 +1,687 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env python3
|
| 2 |
+
# Portions Copyright (c) Meta Platforms, Inc. and affiliates.
|
| 3 |
+
# All rights reserved.
|
| 4 |
+
|
| 5 |
+
# This source code is licensed under the license found in the
|
| 6 |
+
# LICENSE file in the root directory of this source tree.
|
| 7 |
+
|
| 8 |
+
import gzip
|
| 9 |
+
import html
|
| 10 |
+
import io
|
| 11 |
+
import math
|
| 12 |
+
from functools import lru_cache
|
| 13 |
+
from typing import Callable, List, Optional
|
| 14 |
+
|
| 15 |
+
import ftfy
|
| 16 |
+
|
| 17 |
+
import numpy as np
|
| 18 |
+
import regex as re
|
| 19 |
+
import torch
|
| 20 |
+
import torch.nn as nn
|
| 21 |
+
from iopath.common.file_io import g_pathmgr
|
| 22 |
+
from timm.models.layers import trunc_normal_
|
| 23 |
+
|
| 24 |
+
from models.helpers import cast_if_src_dtype, VerboseNNModule
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
def get_sinusoid_encoding_table(n_position, d_hid):
|
| 28 |
+
"""Sinusoid position encoding table"""
|
| 29 |
+
|
| 30 |
+
# TODO: make it with torch instead of numpy
|
| 31 |
+
def get_position_angle_vec(position):
|
| 32 |
+
return [
|
| 33 |
+
position / np.power(10000, 2 * (hid_j // 2) / d_hid)
|
| 34 |
+
for hid_j in range(d_hid)
|
| 35 |
+
]
|
| 36 |
+
|
| 37 |
+
sinusoid_table = np.array(
|
| 38 |
+
[get_position_angle_vec(pos_i) for pos_i in range(n_position)]
|
| 39 |
+
)
|
| 40 |
+
sinusoid_table[:, 0::2] = np.sin(sinusoid_table[:, 0::2]) # dim 2i
|
| 41 |
+
sinusoid_table[:, 1::2] = np.cos(sinusoid_table[:, 1::2]) # dim 2i+1
|
| 42 |
+
|
| 43 |
+
return torch.FloatTensor(sinusoid_table).unsqueeze(0)
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
def interpolate_pos_encoding_2d(target_spatial_size, pos_embed):
|
| 47 |
+
N = pos_embed.shape[1]
|
| 48 |
+
if N == target_spatial_size:
|
| 49 |
+
return pos_embed
|
| 50 |
+
dim = pos_embed.shape[-1]
|
| 51 |
+
# nn.functional.interpolate doesn't work with bfloat16 so we cast to float32
|
| 52 |
+
pos_embed, updated = cast_if_src_dtype(pos_embed, torch.bfloat16, torch.float32)
|
| 53 |
+
pos_embed = nn.functional.interpolate(
|
| 54 |
+
pos_embed.reshape(1, int(math.sqrt(N)), int(math.sqrt(N)), dim).permute(
|
| 55 |
+
0, 3, 1, 2
|
| 56 |
+
),
|
| 57 |
+
scale_factor=math.sqrt(target_spatial_size / N),
|
| 58 |
+
mode="bicubic",
|
| 59 |
+
)
|
| 60 |
+
if updated:
|
| 61 |
+
pos_embed, _ = cast_if_src_dtype(pos_embed, torch.float32, torch.bfloat16)
|
| 62 |
+
pos_embed = pos_embed.permute(0, 2, 3, 1).view(1, -1, dim)
|
| 63 |
+
return pos_embed
|
| 64 |
+
|
| 65 |
+
|
| 66 |
+
def interpolate_pos_encoding(
|
| 67 |
+
npatch_per_img,
|
| 68 |
+
pos_embed,
|
| 69 |
+
patches_layout,
|
| 70 |
+
input_shape=None,
|
| 71 |
+
first_patch_idx=1,
|
| 72 |
+
):
|
| 73 |
+
assert first_patch_idx == 0 or first_patch_idx == 1, "there is 1 CLS token or none"
|
| 74 |
+
N = pos_embed.shape[1] - first_patch_idx # since it's 1 if cls_token exists
|
| 75 |
+
if npatch_per_img == N:
|
| 76 |
+
return pos_embed
|
| 77 |
+
|
| 78 |
+
assert (
|
| 79 |
+
patches_layout[-1] == patches_layout[-2]
|
| 80 |
+
), "Interpolation of pos embed not supported for non-square layouts"
|
| 81 |
+
|
| 82 |
+
class_emb = pos_embed[:, :first_patch_idx]
|
| 83 |
+
pos_embed = pos_embed[:, first_patch_idx:]
|
| 84 |
+
|
| 85 |
+
if input_shape is None or patches_layout[0] == 1:
|
| 86 |
+
# simple 2D pos embedding, no temporal component
|
| 87 |
+
pos_embed = interpolate_pos_encoding_2d(npatch_per_img, pos_embed)
|
| 88 |
+
elif patches_layout[0] > 1:
|
| 89 |
+
# pos embed has a temporal component
|
| 90 |
+
assert len(input_shape) == 4, "temporal interpolation not supported"
|
| 91 |
+
# we only support 2D interpolation in this case
|
| 92 |
+
num_frames = patches_layout[0]
|
| 93 |
+
num_spatial_tokens = patches_layout[1] * patches_layout[2]
|
| 94 |
+
pos_embed = pos_embed.view(1, num_frames, num_spatial_tokens, -1)
|
| 95 |
+
# interpolate embedding for zeroth frame
|
| 96 |
+
pos_embed = interpolate_pos_encoding_2d(
|
| 97 |
+
npatch_per_img, pos_embed[0, 0, ...].unsqueeze(0)
|
| 98 |
+
)
|
| 99 |
+
else:
|
| 100 |
+
raise ValueError("This type of interpolation isn't implemented")
|
| 101 |
+
|
| 102 |
+
return torch.cat((class_emb, pos_embed), dim=1)
|
| 103 |
+
|
| 104 |
+
|
| 105 |
+
def _get_pos_embedding(
|
| 106 |
+
npatch_per_img,
|
| 107 |
+
pos_embed,
|
| 108 |
+
patches_layout,
|
| 109 |
+
input_shape,
|
| 110 |
+
first_patch_idx=1,
|
| 111 |
+
):
|
| 112 |
+
pos_embed = interpolate_pos_encoding(
|
| 113 |
+
npatch_per_img,
|
| 114 |
+
pos_embed,
|
| 115 |
+
patches_layout,
|
| 116 |
+
input_shape=input_shape,
|
| 117 |
+
first_patch_idx=first_patch_idx,
|
| 118 |
+
)
|
| 119 |
+
return pos_embed
|
| 120 |
+
|
| 121 |
+
|
| 122 |
+
class PatchEmbedGeneric(nn.Module):
|
| 123 |
+
"""
|
| 124 |
+
PatchEmbed from Hydra
|
| 125 |
+
"""
|
| 126 |
+
|
| 127 |
+
def __init__(self, proj_stem, norm_layer: Optional[nn.Module] = None):
|
| 128 |
+
super().__init__()
|
| 129 |
+
|
| 130 |
+
if len(proj_stem) > 1:
|
| 131 |
+
self.proj = nn.Sequential(*proj_stem)
|
| 132 |
+
else:
|
| 133 |
+
# Special case to be able to load pre-trained models that were
|
| 134 |
+
# trained with a standard stem
|
| 135 |
+
self.proj = proj_stem[0]
|
| 136 |
+
self.norm_layer = norm_layer
|
| 137 |
+
|
| 138 |
+
def get_patch_layout(self, img_size):
|
| 139 |
+
with torch.no_grad():
|
| 140 |
+
dummy_img = torch.zeros(
|
| 141 |
+
[
|
| 142 |
+
1,
|
| 143 |
+
]
|
| 144 |
+
+ img_size
|
| 145 |
+
)
|
| 146 |
+
dummy_out = self.proj(dummy_img)
|
| 147 |
+
embed_dim = dummy_out.shape[1]
|
| 148 |
+
patches_layout = tuple(dummy_out.shape[2:])
|
| 149 |
+
num_patches = np.prod(patches_layout)
|
| 150 |
+
return patches_layout, num_patches, embed_dim
|
| 151 |
+
|
| 152 |
+
def forward(self, x):
|
| 153 |
+
x = self.proj(x)
|
| 154 |
+
# B C (T) H W -> B (T)HW C
|
| 155 |
+
x = x.flatten(2).transpose(1, 2)
|
| 156 |
+
if self.norm_layer is not None:
|
| 157 |
+
x = self.norm_layer(x)
|
| 158 |
+
return x
|
| 159 |
+
|
| 160 |
+
|
| 161 |
+
class SpatioTemporalPosEmbeddingHelper(VerboseNNModule):
|
| 162 |
+
def __init__(
|
| 163 |
+
self,
|
| 164 |
+
patches_layout: List,
|
| 165 |
+
num_patches: int,
|
| 166 |
+
num_cls_tokens: int,
|
| 167 |
+
embed_dim: int,
|
| 168 |
+
learnable: bool,
|
| 169 |
+
) -> None:
|
| 170 |
+
super().__init__()
|
| 171 |
+
self.num_cls_tokens = num_cls_tokens
|
| 172 |
+
self.patches_layout = patches_layout
|
| 173 |
+
self.num_patches = num_patches
|
| 174 |
+
self.num_tokens = num_cls_tokens + num_patches
|
| 175 |
+
self.learnable = learnable
|
| 176 |
+
if self.learnable:
|
| 177 |
+
self.pos_embed = nn.Parameter(torch.zeros(1, self.num_tokens, embed_dim))
|
| 178 |
+
trunc_normal_(self.pos_embed, std=0.02)
|
| 179 |
+
else:
|
| 180 |
+
self.register_buffer(
|
| 181 |
+
"pos_embed", get_sinusoid_encoding_table(self.num_tokens, embed_dim)
|
| 182 |
+
)
|
| 183 |
+
|
| 184 |
+
def get_pos_embedding(self, vision_input, all_vision_tokens):
|
| 185 |
+
input_shape = vision_input.shape
|
| 186 |
+
pos_embed = _get_pos_embedding(
|
| 187 |
+
all_vision_tokens.size(1) - self.num_cls_tokens,
|
| 188 |
+
pos_embed=self.pos_embed,
|
| 189 |
+
patches_layout=self.patches_layout,
|
| 190 |
+
input_shape=input_shape,
|
| 191 |
+
first_patch_idx=self.num_cls_tokens,
|
| 192 |
+
)
|
| 193 |
+
return pos_embed
|
| 194 |
+
|
| 195 |
+
|
| 196 |
+
class RGBDTPreprocessor(VerboseNNModule):
|
| 197 |
+
def __init__(
|
| 198 |
+
self,
|
| 199 |
+
rgbt_stem: PatchEmbedGeneric,
|
| 200 |
+
depth_stem: PatchEmbedGeneric,
|
| 201 |
+
img_size: List = (3, 224, 224),
|
| 202 |
+
num_cls_tokens: int = 1,
|
| 203 |
+
pos_embed_fn: Callable = None,
|
| 204 |
+
use_type_embed: bool = False,
|
| 205 |
+
init_param_style: str = "openclip",
|
| 206 |
+
) -> None:
|
| 207 |
+
super().__init__()
|
| 208 |
+
stem = rgbt_stem if rgbt_stem is not None else depth_stem
|
| 209 |
+
(
|
| 210 |
+
self.patches_layout,
|
| 211 |
+
self.num_patches,
|
| 212 |
+
self.embed_dim,
|
| 213 |
+
) = stem.get_patch_layout(img_size)
|
| 214 |
+
self.rgbt_stem = rgbt_stem
|
| 215 |
+
self.depth_stem = depth_stem
|
| 216 |
+
self.use_pos_embed = pos_embed_fn is not None
|
| 217 |
+
self.use_type_embed = use_type_embed
|
| 218 |
+
self.num_cls_tokens = num_cls_tokens
|
| 219 |
+
|
| 220 |
+
if self.use_pos_embed:
|
| 221 |
+
self.pos_embedding_helper = pos_embed_fn(
|
| 222 |
+
patches_layout=self.patches_layout,
|
| 223 |
+
num_cls_tokens=num_cls_tokens,
|
| 224 |
+
num_patches=self.num_patches,
|
| 225 |
+
embed_dim=self.embed_dim,
|
| 226 |
+
)
|
| 227 |
+
if self.num_cls_tokens > 0:
|
| 228 |
+
self.cls_token = nn.Parameter(
|
| 229 |
+
torch.zeros(1, self.num_cls_tokens, self.embed_dim)
|
| 230 |
+
)
|
| 231 |
+
if self.use_type_embed:
|
| 232 |
+
self.type_embed = nn.Parameter(torch.zeros(1, 1, self.embed_dim))
|
| 233 |
+
|
| 234 |
+
self.init_parameters(init_param_style)
|
| 235 |
+
|
| 236 |
+
@torch.no_grad()
|
| 237 |
+
def init_parameters(self, init_param_style):
|
| 238 |
+
if init_param_style == "openclip":
|
| 239 |
+
# OpenCLIP style initialization
|
| 240 |
+
scale = self.embed_dim**-0.5
|
| 241 |
+
if self.use_pos_embed:
|
| 242 |
+
nn.init.normal_(self.pos_embedding_helper.pos_embed)
|
| 243 |
+
self.pos_embedding_helper.pos_embed *= scale
|
| 244 |
+
|
| 245 |
+
if self.num_cls_tokens > 0:
|
| 246 |
+
nn.init.normal_(self.cls_token)
|
| 247 |
+
self.cls_token *= scale
|
| 248 |
+
elif init_param_style == "vit":
|
| 249 |
+
self.cls_token.data.fill_(0)
|
| 250 |
+
else:
|
| 251 |
+
raise ValueError(f"Unknown init {init_param_style}")
|
| 252 |
+
|
| 253 |
+
if self.use_type_embed:
|
| 254 |
+
nn.init.normal_(self.type_embed)
|
| 255 |
+
|
| 256 |
+
def tokenize_input_and_cls_pos(self, input, stem, mask):
|
| 257 |
+
# tokens is of shape B x L x D
|
| 258 |
+
tokens = stem(input)
|
| 259 |
+
assert tokens.ndim == 3
|
| 260 |
+
assert tokens.shape[2] == self.embed_dim
|
| 261 |
+
B = tokens.shape[0]
|
| 262 |
+
if self.num_cls_tokens > 0:
|
| 263 |
+
class_tokens = self.cls_token.expand(
|
| 264 |
+
B, -1, -1
|
| 265 |
+
) # stole class_tokens impl from Phil Wang, thanks
|
| 266 |
+
tokens = torch.cat((class_tokens, tokens), dim=1)
|
| 267 |
+
if self.use_pos_embed:
|
| 268 |
+
pos_embed = self.pos_embedding_helper.get_pos_embedding(input, tokens)
|
| 269 |
+
tokens = tokens + pos_embed
|
| 270 |
+
if self.use_type_embed:
|
| 271 |
+
tokens = tokens + self.type_embed.expand(B, -1, -1)
|
| 272 |
+
return tokens
|
| 273 |
+
|
| 274 |
+
def forward(self, vision=None, depth=None, patch_mask=None):
|
| 275 |
+
if patch_mask is not None:
|
| 276 |
+
raise NotImplementedError()
|
| 277 |
+
|
| 278 |
+
if vision is not None:
|
| 279 |
+
vision_tokens = self.tokenize_input_and_cls_pos(
|
| 280 |
+
vision, self.rgbt_stem, patch_mask
|
| 281 |
+
)
|
| 282 |
+
|
| 283 |
+
if depth is not None:
|
| 284 |
+
depth_tokens = self.tokenize_input_and_cls_pos(
|
| 285 |
+
depth, self.depth_stem, patch_mask
|
| 286 |
+
)
|
| 287 |
+
|
| 288 |
+
# aggregate tokens
|
| 289 |
+
if vision is not None and depth is not None:
|
| 290 |
+
final_tokens = vision_tokens + depth_tokens
|
| 291 |
+
else:
|
| 292 |
+
final_tokens = vision_tokens if vision is not None else depth_tokens
|
| 293 |
+
return_dict = {
|
| 294 |
+
"trunk": {
|
| 295 |
+
"tokens": final_tokens,
|
| 296 |
+
},
|
| 297 |
+
"head": {},
|
| 298 |
+
}
|
| 299 |
+
return return_dict
|
| 300 |
+
|
| 301 |
+
|
| 302 |
+
class AudioPreprocessor(RGBDTPreprocessor):
|
| 303 |
+
def __init__(self, audio_stem: PatchEmbedGeneric, **kwargs) -> None:
|
| 304 |
+
super().__init__(rgbt_stem=audio_stem, depth_stem=None, **kwargs)
|
| 305 |
+
|
| 306 |
+
def forward(self, audio=None):
|
| 307 |
+
return super().forward(vision=audio)
|
| 308 |
+
|
| 309 |
+
|
| 310 |
+
class ThermalPreprocessor(RGBDTPreprocessor):
|
| 311 |
+
def __init__(self, thermal_stem: PatchEmbedGeneric, **kwargs) -> None:
|
| 312 |
+
super().__init__(rgbt_stem=thermal_stem, depth_stem=None, **kwargs)
|
| 313 |
+
|
| 314 |
+
def forward(self, thermal=None):
|
| 315 |
+
return super().forward(vision=thermal)
|
| 316 |
+
|
| 317 |
+
|
| 318 |
+
def build_causal_attention_mask(context_length):
|
| 319 |
+
# lazily create causal attention mask, with full attention between the vision tokens
|
| 320 |
+
# pytorch uses additive attention mask; fill with -inf
|
| 321 |
+
mask = torch.empty(context_length, context_length, requires_grad=False)
|
| 322 |
+
mask.fill_(float("-inf"))
|
| 323 |
+
mask.triu_(1) # zero out the lower diagonal
|
| 324 |
+
return mask
|
| 325 |
+
|
| 326 |
+
|
| 327 |
+
class TextPreprocessor(VerboseNNModule):
|
| 328 |
+
def __init__(
|
| 329 |
+
self,
|
| 330 |
+
vocab_size: int,
|
| 331 |
+
context_length: int,
|
| 332 |
+
embed_dim: int,
|
| 333 |
+
causal_masking: bool,
|
| 334 |
+
supply_seq_len_to_head: bool = True,
|
| 335 |
+
num_cls_tokens: int = 0,
|
| 336 |
+
init_param_style: str = "openclip",
|
| 337 |
+
) -> None:
|
| 338 |
+
super().__init__()
|
| 339 |
+
self.vocab_size = vocab_size
|
| 340 |
+
self.context_length = context_length
|
| 341 |
+
self.token_embedding = nn.Embedding(vocab_size, embed_dim)
|
| 342 |
+
self.pos_embed = nn.Parameter(
|
| 343 |
+
torch.empty(1, self.context_length + num_cls_tokens, embed_dim)
|
| 344 |
+
)
|
| 345 |
+
self.causal_masking = causal_masking
|
| 346 |
+
if self.causal_masking:
|
| 347 |
+
mask = build_causal_attention_mask(self.context_length)
|
| 348 |
+
# register the mask as a buffer so it can be moved to the right device
|
| 349 |
+
self.register_buffer("mask", mask)
|
| 350 |
+
|
| 351 |
+
self.supply_seq_len_to_head = supply_seq_len_to_head
|
| 352 |
+
self.num_cls_tokens = num_cls_tokens
|
| 353 |
+
self.embed_dim = embed_dim
|
| 354 |
+
if num_cls_tokens > 0:
|
| 355 |
+
assert self.causal_masking is False, "Masking + CLS token isn't implemented"
|
| 356 |
+
self.cls_token = nn.Parameter(
|
| 357 |
+
torch.zeros(1, self.num_cls_tokens, embed_dim)
|
| 358 |
+
)
|
| 359 |
+
|
| 360 |
+
self.init_parameters(init_param_style)
|
| 361 |
+
|
| 362 |
+
@torch.no_grad()
|
| 363 |
+
def init_parameters(self, init_param_style="openclip"):
|
| 364 |
+
# OpenCLIP style initialization
|
| 365 |
+
nn.init.normal_(self.token_embedding.weight, std=0.02)
|
| 366 |
+
nn.init.normal_(self.pos_embed, std=0.01)
|
| 367 |
+
|
| 368 |
+
if init_param_style == "openclip":
|
| 369 |
+
# OpenCLIP style initialization
|
| 370 |
+
scale = self.embed_dim**-0.5
|
| 371 |
+
if self.num_cls_tokens > 0:
|
| 372 |
+
nn.init.normal_(self.cls_token)
|
| 373 |
+
self.cls_token *= scale
|
| 374 |
+
elif init_param_style == "vit":
|
| 375 |
+
self.cls_token.data.fill_(0)
|
| 376 |
+
else:
|
| 377 |
+
raise ValueError(f"Unknown init {init_param_style}")
|
| 378 |
+
|
| 379 |
+
def forward(self, text):
|
| 380 |
+
# text tokens are of shape B x L x D
|
| 381 |
+
text_tokens = self.token_embedding(text)
|
| 382 |
+
# concat CLS tokens if any
|
| 383 |
+
if self.num_cls_tokens > 0:
|
| 384 |
+
B = text_tokens.shape[0]
|
| 385 |
+
class_tokens = self.cls_token.expand(
|
| 386 |
+
B, -1, -1
|
| 387 |
+
) # stole class_tokens impl from Phil Wang, thanks
|
| 388 |
+
text_tokens = torch.cat((class_tokens, text_tokens), dim=1)
|
| 389 |
+
text_tokens = text_tokens + self.pos_embed
|
| 390 |
+
return_dict = {
|
| 391 |
+
"trunk": {
|
| 392 |
+
"tokens": text_tokens,
|
| 393 |
+
},
|
| 394 |
+
"head": {},
|
| 395 |
+
}
|
| 396 |
+
# Compute sequence length after adding CLS tokens
|
| 397 |
+
if self.supply_seq_len_to_head:
|
| 398 |
+
text_lengths = text.argmax(dim=-1)
|
| 399 |
+
return_dict["head"] = {
|
| 400 |
+
"seq_len": text_lengths,
|
| 401 |
+
}
|
| 402 |
+
if self.causal_masking:
|
| 403 |
+
return_dict["trunk"].update({"attn_mask": self.mask})
|
| 404 |
+
return return_dict
|
| 405 |
+
|
| 406 |
+
|
| 407 |
+
class Im2Video(nn.Module):
|
| 408 |
+
"""Convert an image into a trivial video."""
|
| 409 |
+
|
| 410 |
+
def __init__(self, time_dim=2):
|
| 411 |
+
super().__init__()
|
| 412 |
+
self.time_dim = time_dim
|
| 413 |
+
|
| 414 |
+
def forward(self, x):
|
| 415 |
+
if x.ndim == 4:
|
| 416 |
+
# B, C, H, W -> B, C, T, H, W
|
| 417 |
+
return x.unsqueeze(self.time_dim)
|
| 418 |
+
elif x.ndim == 5:
|
| 419 |
+
return x
|
| 420 |
+
else:
|
| 421 |
+
raise ValueError(f"Dimension incorrect {x.shape}")
|
| 422 |
+
|
| 423 |
+
|
| 424 |
+
class PadIm2Video(Im2Video):
|
| 425 |
+
def __init__(self, ntimes, pad_type, time_dim=2):
|
| 426 |
+
super().__init__(time_dim=time_dim)
|
| 427 |
+
assert ntimes > 0
|
| 428 |
+
assert pad_type in ["zero", "repeat"]
|
| 429 |
+
self.ntimes = ntimes
|
| 430 |
+
self.pad_type = pad_type
|
| 431 |
+
|
| 432 |
+
def forward(self, x):
|
| 433 |
+
x = super().forward(x)
|
| 434 |
+
if x.shape[self.time_dim] == 1:
|
| 435 |
+
if self.pad_type == "repeat":
|
| 436 |
+
new_shape = [1] * len(x.shape)
|
| 437 |
+
new_shape[self.time_dim] = self.ntimes
|
| 438 |
+
x = x.repeat(new_shape)
|
| 439 |
+
elif self.pad_type == "zero":
|
| 440 |
+
padarg = [0, 0] * len(x.shape)
|
| 441 |
+
padarg[2 * self.time_dim + 1] = self.ntimes - x.shape[self.time_dim]
|
| 442 |
+
x = nn.functional.pad(x, padarg)
|
| 443 |
+
return x
|
| 444 |
+
|
| 445 |
+
|
| 446 |
+
# Modified from github.com/openai/CLIP
|
| 447 |
+
@lru_cache()
|
| 448 |
+
def bytes_to_unicode():
|
| 449 |
+
"""
|
| 450 |
+
Returns list of utf-8 byte and a corresponding list of unicode strings.
|
| 451 |
+
The reversible bpe codes work on unicode strings.
|
| 452 |
+
This means you need a large # of unicode characters in your vocab if you want to avoid UNKs.
|
| 453 |
+
When you're at something like a 10B token dataset you end up needing around 5K for decent coverage.
|
| 454 |
+
This is a signficant percentage of your normal, say, 32K bpe vocab.
|
| 455 |
+
To avoid that, we want lookup tables between utf-8 bytes and unicode strings.
|
| 456 |
+
And avoids mapping to whitespace/control characters the bpe code barfs on.
|
| 457 |
+
"""
|
| 458 |
+
bs = (
|
| 459 |
+
list(range(ord("!"), ord("~") + 1))
|
| 460 |
+
+ list(range(ord("¡"), ord("¬") + 1))
|
| 461 |
+
+ list(range(ord("®"), ord("ÿ") + 1))
|
| 462 |
+
)
|
| 463 |
+
cs = bs[:]
|
| 464 |
+
n = 0
|
| 465 |
+
for b in range(2**8):
|
| 466 |
+
if b not in bs:
|
| 467 |
+
bs.append(b)
|
| 468 |
+
cs.append(2**8 + n)
|
| 469 |
+
n += 1
|
| 470 |
+
cs = [chr(n) for n in cs]
|
| 471 |
+
return dict(zip(bs, cs))
|
| 472 |
+
|
| 473 |
+
|
| 474 |
+
def get_pairs(word):
|
| 475 |
+
"""Return set of symbol pairs in a word.
|
| 476 |
+
Word is represented as tuple of symbols (symbols being variable-length strings).
|
| 477 |
+
"""
|
| 478 |
+
pairs = set()
|
| 479 |
+
prev_char = word[0]
|
| 480 |
+
for char in word[1:]:
|
| 481 |
+
pairs.add((prev_char, char))
|
| 482 |
+
prev_char = char
|
| 483 |
+
return pairs
|
| 484 |
+
|
| 485 |
+
|
| 486 |
+
def basic_clean(text):
|
| 487 |
+
text = ftfy.fix_text(text)
|
| 488 |
+
text = html.unescape(html.unescape(text))
|
| 489 |
+
return text.strip()
|
| 490 |
+
|
| 491 |
+
|
| 492 |
+
def whitespace_clean(text):
|
| 493 |
+
text = re.sub(r"\s+", " ", text)
|
| 494 |
+
text = text.strip()
|
| 495 |
+
return text
|
| 496 |
+
|
| 497 |
+
|
| 498 |
+
class SimpleTokenizer(object):
|
| 499 |
+
def __init__(self, bpe_path: str, context_length=77):
|
| 500 |
+
self.byte_encoder = bytes_to_unicode()
|
| 501 |
+
self.byte_decoder = {v: k for k, v in self.byte_encoder.items()}
|
| 502 |
+
|
| 503 |
+
with g_pathmgr.open(bpe_path, "rb") as fh:
|
| 504 |
+
bpe_bytes = io.BytesIO(fh.read())
|
| 505 |
+
merges = gzip.open(bpe_bytes).read().decode("utf-8").split("\n")
|
| 506 |
+
merges = merges[1 : 49152 - 256 - 2 + 1]
|
| 507 |
+
merges = [tuple(merge.split()) for merge in merges]
|
| 508 |
+
vocab = list(bytes_to_unicode().values())
|
| 509 |
+
vocab = vocab + [v + "</w>" for v in vocab]
|
| 510 |
+
for merge in merges:
|
| 511 |
+
vocab.append("".join(merge))
|
| 512 |
+
vocab.extend(["<|startoftext|>", "<|endoftext|>"])
|
| 513 |
+
self.encoder = dict(zip(vocab, range(len(vocab))))
|
| 514 |
+
self.decoder = {v: k for k, v in self.encoder.items()}
|
| 515 |
+
self.bpe_ranks = dict(zip(merges, range(len(merges))))
|
| 516 |
+
self.cache = {
|
| 517 |
+
"<|startoftext|>": "<|startoftext|>",
|
| 518 |
+
"<|endoftext|>": "<|endoftext|>",
|
| 519 |
+
}
|
| 520 |
+
self.pat = re.compile(
|
| 521 |
+
r"""<\|startoftext\|>|<\|endoftext\|>|'s|'t|'re|'ve|'m|'ll|'d|[\p{L}]+|[\p{N}]|[^\s\p{L}\p{N}]+""",
|
| 522 |
+
re.IGNORECASE,
|
| 523 |
+
)
|
| 524 |
+
self.context_length = context_length
|
| 525 |
+
|
| 526 |
+
def bpe(self, token):
|
| 527 |
+
if token in self.cache:
|
| 528 |
+
return self.cache[token]
|
| 529 |
+
word = tuple(token[:-1]) + (token[-1] + "</w>",)
|
| 530 |
+
pairs = get_pairs(word)
|
| 531 |
+
|
| 532 |
+
if not pairs:
|
| 533 |
+
return token + "</w>"
|
| 534 |
+
|
| 535 |
+
while True:
|
| 536 |
+
bigram = min(pairs, key=lambda pair: self.bpe_ranks.get(pair, float("inf")))
|
| 537 |
+
if bigram not in self.bpe_ranks:
|
| 538 |
+
break
|
| 539 |
+
first, second = bigram
|
| 540 |
+
new_word = []
|
| 541 |
+
i = 0
|
| 542 |
+
while i < len(word):
|
| 543 |
+
try:
|
| 544 |
+
j = word.index(first, i)
|
| 545 |
+
new_word.extend(word[i:j])
|
| 546 |
+
i = j
|
| 547 |
+
except:
|
| 548 |
+
new_word.extend(word[i:])
|
| 549 |
+
break
|
| 550 |
+
|
| 551 |
+
if word[i] == first and i < len(word) - 1 and word[i + 1] == second:
|
| 552 |
+
new_word.append(first + second)
|
| 553 |
+
i += 2
|
| 554 |
+
else:
|
| 555 |
+
new_word.append(word[i])
|
| 556 |
+
i += 1
|
| 557 |
+
new_word = tuple(new_word)
|
| 558 |
+
word = new_word
|
| 559 |
+
if len(word) == 1:
|
| 560 |
+
break
|
| 561 |
+
else:
|
| 562 |
+
pairs = get_pairs(word)
|
| 563 |
+
word = " ".join(word)
|
| 564 |
+
self.cache[token] = word
|
| 565 |
+
return word
|
| 566 |
+
|
| 567 |
+
def encode(self, text):
|
| 568 |
+
bpe_tokens = []
|
| 569 |
+
text = whitespace_clean(basic_clean(text)).lower()
|
| 570 |
+
for token in re.findall(self.pat, text):
|
| 571 |
+
token = "".join(self.byte_encoder[b] for b in token.encode("utf-8"))
|
| 572 |
+
bpe_tokens.extend(
|
| 573 |
+
self.encoder[bpe_token] for bpe_token in self.bpe(token).split(" ")
|
| 574 |
+
)
|
| 575 |
+
return bpe_tokens
|
| 576 |
+
|
| 577 |
+
def decode(self, tokens):
|
| 578 |
+
text = "".join([self.decoder[token] for token in tokens])
|
| 579 |
+
text = (
|
| 580 |
+
bytearray([self.byte_decoder[c] for c in text])
|
| 581 |
+
.decode("utf-8", errors="replace")
|
| 582 |
+
.replace("</w>", " ")
|
| 583 |
+
)
|
| 584 |
+
return text
|
| 585 |
+
|
| 586 |
+
def __call__(self, texts, context_length=None):
|
| 587 |
+
if not context_length:
|
| 588 |
+
context_length = self.context_length
|
| 589 |
+
|
| 590 |
+
if isinstance(texts, str):
|
| 591 |
+
texts = [texts]
|
| 592 |
+
|
| 593 |
+
sot_token = self.encoder["<|startoftext|>"]
|
| 594 |
+
eot_token = self.encoder["<|endoftext|>"]
|
| 595 |
+
all_tokens = [[sot_token] + self.encode(text) + [eot_token] for text in texts]
|
| 596 |
+
result = torch.zeros(len(all_tokens), context_length, dtype=torch.long)
|
| 597 |
+
|
| 598 |
+
for i, tokens in enumerate(all_tokens):
|
| 599 |
+
tokens = tokens[:context_length]
|
| 600 |
+
result[i, : len(tokens)] = torch.tensor(tokens)
|
| 601 |
+
|
| 602 |
+
if len(result) == 1:
|
| 603 |
+
return result[0]
|
| 604 |
+
return result
|
| 605 |
+
|
| 606 |
+
|
| 607 |
+
class IMUPreprocessor(VerboseNNModule):
|
| 608 |
+
def __init__(
|
| 609 |
+
self,
|
| 610 |
+
kernel_size: int,
|
| 611 |
+
imu_stem: PatchEmbedGeneric,
|
| 612 |
+
embed_dim: int,
|
| 613 |
+
img_size: List = (6, 2000),
|
| 614 |
+
num_cls_tokens: int = 1,
|
| 615 |
+
pos_embed_fn: Callable = None,
|
| 616 |
+
init_param_style: str = "openclip",
|
| 617 |
+
) -> None:
|
| 618 |
+
super().__init__()
|
| 619 |
+
stem = imu_stem
|
| 620 |
+
self.imu_stem = imu_stem
|
| 621 |
+
self.embed_dim = embed_dim
|
| 622 |
+
self.use_pos_embed = pos_embed_fn is not None
|
| 623 |
+
self.num_cls_tokens = num_cls_tokens
|
| 624 |
+
self.kernel_size = kernel_size
|
| 625 |
+
self.pos_embed = nn.Parameter(
|
| 626 |
+
torch.empty(1, (img_size[1] // kernel_size) + num_cls_tokens, embed_dim)
|
| 627 |
+
)
|
| 628 |
+
|
| 629 |
+
if self.num_cls_tokens > 0:
|
| 630 |
+
self.cls_token = nn.Parameter(
|
| 631 |
+
torch.zeros(1, self.num_cls_tokens, self.embed_dim)
|
| 632 |
+
)
|
| 633 |
+
|
| 634 |
+
self.init_parameters(init_param_style)
|
| 635 |
+
|
| 636 |
+
@torch.no_grad()
|
| 637 |
+
def init_parameters(self, init_param_style):
|
| 638 |
+
nn.init.normal_(self.pos_embed, std=0.01)
|
| 639 |
+
|
| 640 |
+
if init_param_style == "openclip":
|
| 641 |
+
# OpenCLIP style initialization
|
| 642 |
+
scale = self.embed_dim**-0.5
|
| 643 |
+
|
| 644 |
+
if self.num_cls_tokens > 0:
|
| 645 |
+
nn.init.normal_(self.cls_token)
|
| 646 |
+
self.cls_token *= scale
|
| 647 |
+
elif init_param_style == "vit":
|
| 648 |
+
self.cls_token.data.fill_(0)
|
| 649 |
+
else:
|
| 650 |
+
raise ValueError(f"Unknown init {init_param_style}")
|
| 651 |
+
|
| 652 |
+
def tokenize_input_and_cls_pos(self, input, stem):
|
| 653 |
+
# tokens is of shape B x L x D
|
| 654 |
+
tokens = stem.norm_layer(stem.proj(input))
|
| 655 |
+
assert tokens.ndim == 3
|
| 656 |
+
assert tokens.shape[2] == self.embed_dim
|
| 657 |
+
B = tokens.shape[0]
|
| 658 |
+
if self.num_cls_tokens > 0:
|
| 659 |
+
class_tokens = self.cls_token.expand(
|
| 660 |
+
B, -1, -1
|
| 661 |
+
) # stole class_tokens impl from Phil Wang, thanks
|
| 662 |
+
tokens = torch.cat((class_tokens, tokens), dim=1)
|
| 663 |
+
if self.use_pos_embed:
|
| 664 |
+
tokens = tokens + self.pos_embed
|
| 665 |
+
return tokens
|
| 666 |
+
|
| 667 |
+
def forward(self, imu):
|
| 668 |
+
# Patchify
|
| 669 |
+
imu = imu.unfold(
|
| 670 |
+
-1,
|
| 671 |
+
self.kernel_size,
|
| 672 |
+
self.kernel_size,
|
| 673 |
+
).permute(0, 2, 1, 3)
|
| 674 |
+
imu = imu.reshape(imu.size(0), imu.size(1), -1)
|
| 675 |
+
|
| 676 |
+
imu_tokens = self.tokenize_input_and_cls_pos(
|
| 677 |
+
imu,
|
| 678 |
+
self.imu_stem,
|
| 679 |
+
)
|
| 680 |
+
|
| 681 |
+
return_dict = {
|
| 682 |
+
"trunk": {
|
| 683 |
+
"tokens": imu_tokens,
|
| 684 |
+
},
|
| 685 |
+
"head": {},
|
| 686 |
+
}
|
| 687 |
+
return return_dict
|
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/transformer.py
ADDED
|
@@ -0,0 +1,284 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env python3
|
| 2 |
+
# Portions Copyright (c) Meta Platforms, Inc. and affiliates.
|
| 3 |
+
# All rights reserved.
|
| 4 |
+
|
| 5 |
+
# This source code is licensed under the license found in the
|
| 6 |
+
# LICENSE file in the root directory of this source tree.
|
| 7 |
+
|
| 8 |
+
# Code modified from
|
| 9 |
+
# https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py ;
|
| 10 |
+
# https://github.com/facebookresearch/deit/blob/main/models.py
|
| 11 |
+
# and https://github.com/facebookresearch/vissl/blob/main/vissl/models/trunks/vision_transformer.py
|
| 12 |
+
|
| 13 |
+
|
| 14 |
+
import copy
|
| 15 |
+
import fnmatch
|
| 16 |
+
import logging
|
| 17 |
+
from functools import partial
|
| 18 |
+
from typing import Callable, List
|
| 19 |
+
|
| 20 |
+
import torch
|
| 21 |
+
import torch.nn as nn
|
| 22 |
+
import torch.utils.checkpoint as checkpoint
|
| 23 |
+
|
| 24 |
+
from timm.models.layers import DropPath, trunc_normal_
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
class Attention(nn.Module):
|
| 28 |
+
def __init__(
|
| 29 |
+
self,
|
| 30 |
+
dim,
|
| 31 |
+
num_heads=8,
|
| 32 |
+
qkv_bias=False,
|
| 33 |
+
qk_scale=None,
|
| 34 |
+
attn_drop=0.0,
|
| 35 |
+
proj_drop=0.0,
|
| 36 |
+
):
|
| 37 |
+
super().__init__()
|
| 38 |
+
self.num_heads = num_heads
|
| 39 |
+
head_dim = dim // num_heads
|
| 40 |
+
# NOTE scale factor was wrong in my original version,
|
| 41 |
+
# can set manually to be compat with prev weights
|
| 42 |
+
self.scale = qk_scale or head_dim**-0.5
|
| 43 |
+
|
| 44 |
+
self.qkv = nn.Linear(dim, dim * 3, bias=qkv_bias)
|
| 45 |
+
self.attn_drop = nn.Dropout(attn_drop)
|
| 46 |
+
self.proj = nn.Linear(dim, dim)
|
| 47 |
+
self.proj_drop = nn.Dropout(proj_drop)
|
| 48 |
+
|
| 49 |
+
def forward(self, x):
|
| 50 |
+
B, N, C = x.shape
|
| 51 |
+
qkv = (
|
| 52 |
+
self.qkv(x)
|
| 53 |
+
.reshape(B, N, 3, self.num_heads, C // self.num_heads)
|
| 54 |
+
.permute(2, 0, 3, 1, 4)
|
| 55 |
+
)
|
| 56 |
+
q, k, v = (
|
| 57 |
+
qkv[0],
|
| 58 |
+
qkv[1],
|
| 59 |
+
qkv[2],
|
| 60 |
+
) # make torchscript happy (cannot use tensor as tuple)
|
| 61 |
+
|
| 62 |
+
attn = (q @ k.transpose(-2, -1)) * self.scale
|
| 63 |
+
attn = attn.softmax(dim=-1)
|
| 64 |
+
attn = self.attn_drop(attn)
|
| 65 |
+
|
| 66 |
+
x = (attn @ v).transpose(1, 2).reshape(B, N, C)
|
| 67 |
+
x = self.proj(x)
|
| 68 |
+
x = self.proj_drop(x)
|
| 69 |
+
return x
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
class Mlp(nn.Module):
|
| 73 |
+
def __init__(
|
| 74 |
+
self,
|
| 75 |
+
in_features,
|
| 76 |
+
hidden_features=None,
|
| 77 |
+
out_features=None,
|
| 78 |
+
act_layer=nn.GELU,
|
| 79 |
+
drop=0.0,
|
| 80 |
+
):
|
| 81 |
+
super().__init__()
|
| 82 |
+
out_features = out_features or in_features
|
| 83 |
+
hidden_features = hidden_features or in_features
|
| 84 |
+
self.fc1 = nn.Linear(in_features, hidden_features)
|
| 85 |
+
self.act = act_layer()
|
| 86 |
+
self.fc2 = nn.Linear(hidden_features, out_features)
|
| 87 |
+
self.drop = nn.Dropout(drop)
|
| 88 |
+
|
| 89 |
+
def forward(self, x):
|
| 90 |
+
x = self.fc1(x)
|
| 91 |
+
x = self.act(x)
|
| 92 |
+
x = self.drop(x)
|
| 93 |
+
x = self.fc2(x)
|
| 94 |
+
x = self.drop(x)
|
| 95 |
+
return x
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
class MultiheadAttention(nn.MultiheadAttention):
|
| 99 |
+
def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
|
| 100 |
+
return super().forward(x, x, x, need_weights=False, attn_mask=attn_mask)[0]
|
| 101 |
+
|
| 102 |
+
|
| 103 |
+
class ViTAttention(Attention):
|
| 104 |
+
def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
|
| 105 |
+
assert attn_mask is None
|
| 106 |
+
return super().forward(x)
|
| 107 |
+
|
| 108 |
+
|
| 109 |
+
class BlockWithMasking(nn.Module):
|
| 110 |
+
def __init__(
|
| 111 |
+
self,
|
| 112 |
+
dim: int,
|
| 113 |
+
attn_target: Callable,
|
| 114 |
+
mlp_ratio: int = 4,
|
| 115 |
+
act_layer: Callable = nn.GELU,
|
| 116 |
+
norm_layer: Callable = nn.LayerNorm,
|
| 117 |
+
ffn_dropout_rate: float = 0.0,
|
| 118 |
+
drop_path: float = 0.0,
|
| 119 |
+
layer_scale_type: str = None,
|
| 120 |
+
layer_scale_init_value: float = 1e-4,
|
| 121 |
+
):
|
| 122 |
+
super().__init__()
|
| 123 |
+
|
| 124 |
+
assert not isinstance(
|
| 125 |
+
attn_target, nn.Module
|
| 126 |
+
), "attn_target should be a Callable. Otherwise attn_target is shared across blocks!"
|
| 127 |
+
self.attn = attn_target()
|
| 128 |
+
if drop_path > 0.0:
|
| 129 |
+
self.drop_path = DropPath(drop_path)
|
| 130 |
+
else:
|
| 131 |
+
self.drop_path = nn.Identity()
|
| 132 |
+
self.norm_1 = norm_layer(dim)
|
| 133 |
+
mlp_hidden_dim = int(mlp_ratio * dim)
|
| 134 |
+
self.mlp = Mlp(
|
| 135 |
+
in_features=dim,
|
| 136 |
+
hidden_features=mlp_hidden_dim,
|
| 137 |
+
act_layer=act_layer,
|
| 138 |
+
drop=ffn_dropout_rate,
|
| 139 |
+
)
|
| 140 |
+
self.norm_2 = norm_layer(dim)
|
| 141 |
+
self.layer_scale_type = layer_scale_type
|
| 142 |
+
if self.layer_scale_type is not None:
|
| 143 |
+
assert self.layer_scale_type in [
|
| 144 |
+
"per_channel",
|
| 145 |
+
"scalar",
|
| 146 |
+
], f"Found Layer scale type {self.layer_scale_type}"
|
| 147 |
+
if self.layer_scale_type == "per_channel":
|
| 148 |
+
# one gamma value per channel
|
| 149 |
+
gamma_shape = [1, 1, dim]
|
| 150 |
+
elif self.layer_scale_type == "scalar":
|
| 151 |
+
# single gamma value for all channels
|
| 152 |
+
gamma_shape = [1, 1, 1]
|
| 153 |
+
# two gammas: for each part of the fwd in the encoder
|
| 154 |
+
self.layer_scale_gamma1 = nn.Parameter(
|
| 155 |
+
torch.ones(size=gamma_shape) * layer_scale_init_value,
|
| 156 |
+
requires_grad=True,
|
| 157 |
+
)
|
| 158 |
+
self.layer_scale_gamma2 = nn.Parameter(
|
| 159 |
+
torch.ones(size=gamma_shape) * layer_scale_init_value,
|
| 160 |
+
requires_grad=True,
|
| 161 |
+
)
|
| 162 |
+
|
| 163 |
+
def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
|
| 164 |
+
if self.layer_scale_type is None:
|
| 165 |
+
x = x + self.drop_path(self.attn(self.norm_1(x), attn_mask))
|
| 166 |
+
x = x + self.drop_path(self.mlp(self.norm_2(x)))
|
| 167 |
+
else:
|
| 168 |
+
x = (
|
| 169 |
+
x
|
| 170 |
+
+ self.drop_path(self.attn(self.norm_1(x), attn_mask))
|
| 171 |
+
* self.layer_scale_gamma1
|
| 172 |
+
)
|
| 173 |
+
x = x + self.drop_path(self.mlp(self.norm_2(x))) * self.layer_scale_gamma2
|
| 174 |
+
return x
|
| 175 |
+
|
| 176 |
+
|
| 177 |
+
_LAYER_NORM = partial(nn.LayerNorm, eps=1e-6)
|
| 178 |
+
|
| 179 |
+
|
| 180 |
+
class SimpleTransformer(nn.Module):
|
| 181 |
+
def __init__(
|
| 182 |
+
self,
|
| 183 |
+
attn_target: Callable,
|
| 184 |
+
embed_dim: int,
|
| 185 |
+
num_blocks: int,
|
| 186 |
+
block: Callable = BlockWithMasking,
|
| 187 |
+
pre_transformer_layer: Callable = None,
|
| 188 |
+
post_transformer_layer: Callable = None,
|
| 189 |
+
drop_path_rate: float = 0.0,
|
| 190 |
+
drop_path_type: str = "progressive",
|
| 191 |
+
norm_layer: Callable = _LAYER_NORM,
|
| 192 |
+
mlp_ratio: int = 4,
|
| 193 |
+
ffn_dropout_rate: float = 0.0,
|
| 194 |
+
layer_scale_type: str = None, # from cait; possible values are None, "per_channel", "scalar"
|
| 195 |
+
layer_scale_init_value: float = 1e-4, # from cait; float
|
| 196 |
+
weight_init_style: str = "jax", # possible values jax or pytorch
|
| 197 |
+
):
|
| 198 |
+
"""
|
| 199 |
+
Simple Transformer with the following features
|
| 200 |
+
1. Supports masked attention
|
| 201 |
+
2. Supports DropPath
|
| 202 |
+
3. Supports LayerScale
|
| 203 |
+
4. Supports Dropout in Attention and FFN
|
| 204 |
+
5. Makes few assumptions about the input except that it is a Tensor
|
| 205 |
+
"""
|
| 206 |
+
super().__init__()
|
| 207 |
+
self.pre_transformer_layer = pre_transformer_layer
|
| 208 |
+
if drop_path_type == "progressive":
|
| 209 |
+
dpr = [x.item() for x in torch.linspace(0, drop_path_rate, num_blocks)]
|
| 210 |
+
elif drop_path_type == "uniform":
|
| 211 |
+
dpr = [drop_path_rate for i in range(num_blocks)]
|
| 212 |
+
else:
|
| 213 |
+
raise ValueError(f"Unknown drop_path_type: {drop_path_type}")
|
| 214 |
+
|
| 215 |
+
self.blocks = nn.Sequential(
|
| 216 |
+
*[
|
| 217 |
+
block(
|
| 218 |
+
dim=embed_dim,
|
| 219 |
+
attn_target=attn_target,
|
| 220 |
+
mlp_ratio=mlp_ratio,
|
| 221 |
+
ffn_dropout_rate=ffn_dropout_rate,
|
| 222 |
+
drop_path=dpr[i],
|
| 223 |
+
norm_layer=norm_layer,
|
| 224 |
+
layer_scale_type=layer_scale_type,
|
| 225 |
+
layer_scale_init_value=layer_scale_init_value,
|
| 226 |
+
)
|
| 227 |
+
for i in range(num_blocks)
|
| 228 |
+
]
|
| 229 |
+
)
|
| 230 |
+
self.post_transformer_layer = post_transformer_layer
|
| 231 |
+
self.weight_init_style = weight_init_style
|
| 232 |
+
self.apply(self._init_weights)
|
| 233 |
+
|
| 234 |
+
def _init_weights(self, m):
|
| 235 |
+
if isinstance(m, nn.Linear):
|
| 236 |
+
if self.weight_init_style == "jax":
|
| 237 |
+
# Based on MAE and official Jax ViT implementation
|
| 238 |
+
torch.nn.init.xavier_uniform_(m.weight)
|
| 239 |
+
elif self.weight_init_style == "pytorch":
|
| 240 |
+
# PyTorch ViT uses trunc_normal_
|
| 241 |
+
trunc_normal_(m.weight, std=0.02)
|
| 242 |
+
|
| 243 |
+
if m.bias is not None:
|
| 244 |
+
nn.init.constant_(m.bias, 0)
|
| 245 |
+
elif isinstance(m, (nn.LayerNorm)):
|
| 246 |
+
nn.init.constant_(m.bias, 0)
|
| 247 |
+
nn.init.constant_(m.weight, 1.0)
|
| 248 |
+
|
| 249 |
+
def forward(
|
| 250 |
+
self,
|
| 251 |
+
tokens: torch.Tensor,
|
| 252 |
+
attn_mask: torch.Tensor = None,
|
| 253 |
+
use_checkpoint: bool = False,
|
| 254 |
+
checkpoint_every_n: int = 1,
|
| 255 |
+
checkpoint_blk_ids: List[int] = None,
|
| 256 |
+
):
|
| 257 |
+
"""
|
| 258 |
+
Inputs
|
| 259 |
+
- tokens: data of shape N x L x D (or L x N x D depending on the attention implementation)
|
| 260 |
+
- attn: mask of shape L x L
|
| 261 |
+
|
| 262 |
+
Output
|
| 263 |
+
- x: data of shape N x L x D (or L x N x D depending on the attention implementation)
|
| 264 |
+
"""
|
| 265 |
+
if self.pre_transformer_layer:
|
| 266 |
+
tokens = self.pre_transformer_layer(tokens)
|
| 267 |
+
if use_checkpoint and checkpoint_blk_ids is None:
|
| 268 |
+
checkpoint_blk_ids = [
|
| 269 |
+
blk_id
|
| 270 |
+
for blk_id in range(len(self.blocks))
|
| 271 |
+
if blk_id % checkpoint_every_n == 0
|
| 272 |
+
]
|
| 273 |
+
if checkpoint_blk_ids:
|
| 274 |
+
checkpoint_blk_ids = set(checkpoint_blk_ids)
|
| 275 |
+
for blk_id, blk in enumerate(self.blocks):
|
| 276 |
+
if use_checkpoint and blk_id in checkpoint_blk_ids:
|
| 277 |
+
tokens = checkpoint.checkpoint(
|
| 278 |
+
blk, tokens, attn_mask, use_reentrant=False
|
| 279 |
+
)
|
| 280 |
+
else:
|
| 281 |
+
tokens = blk(tokens, attn_mask=attn_mask)
|
| 282 |
+
if self.post_transformer_layer:
|
| 283 |
+
tokens = self.post_transformer_layer(tokens)
|
| 284 |
+
return tokens
|
external/Grounded-Segment-Anything/playground/PaintByExample/README.md
ADDED
|
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
## Paint by Example: Exemplar-based Image Editing with Diffusion Models
|
| 2 |
+
|
| 3 |
+
:grapes: [[Official Project Page](https://github.com/Fantasy-Studio/Paint-by-Example)] :apple:[[Official Online Demo](https://huggingface.co/spaces/Fantasy-Studio/Paint-by-Example)]
|
| 4 |
+
|
| 5 |
+
<div align="center">
|
| 6 |
+
|
| 7 |
+

|
| 8 |
+
|
| 9 |
+
</div>
|
| 10 |
+
|
| 11 |
+
## Abstract
|
| 12 |
+
|
| 13 |
+
> Language-guided image editing has achieved great success recently. In this paper, for the first time, we investigate exemplar-guided image editing for more precise control. We achieve this goal by leveraging self-supervised training to disentangle and re-organize the source image and the exemplar. However, the naive approach will cause obvious fusing artifacts. We carefully analyze it and propose an information bottleneck and strong augmentations to avoid the trivial solution of directly copying and pasting the exemplar image. Meanwhile, to ensure the controllability of the editing process, we design an arbitrary shape mask for the exemplar image and leverage the classifier-free guidance to increase the similarity to the exemplar image. The whole framework involves a single forward of the diffusion model without any iterative optimization. We demonstrate that our method achieves an impressive performance and enables controllable editing on in-the-wild images with high fidelity.
|
| 14 |
+
|
| 15 |
+
## Table of Contents
|
| 16 |
+
- [Installation](#installation)
|
| 17 |
+
- [Paint-By-Example Demos](#paint-by-example-demos)
|
| 18 |
+
- [Diffuser Demo](#paintbyexample-diffuser-demos)
|
| 19 |
+
- [PaintByExample with SAM](#paintbyexample-with-sam)
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
## TODO
|
| 23 |
+
- [x] PaintByExample Diffuser Demo
|
| 24 |
+
- [x] PaintByExample with SAM
|
| 25 |
+
- [ ] PaintByExample with GroundingDINO
|
| 26 |
+
- [ ] PaintByExample with Grounded-SAM
|
| 27 |
+
|
| 28 |
+
## Installation
|
| 29 |
+
We're using PaintByExample with diffusers, install diffusers as follows:
|
| 30 |
+
```bash
|
| 31 |
+
pip install diffusers==0.16.1
|
| 32 |
+
```
|
| 33 |
+
Then install Grounded-SAM follows [Grounded-SAM Installation](https://github.com/IDEA-Research/Grounded-Segment-Anything#installation) for some extension demos.
|
| 34 |
+
|
| 35 |
+
## Paint-By-Example Demos
|
| 36 |
+
Here we provide the demos for `PaintByExample`
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
### PaintByExample Diffuser Demos
|
| 40 |
+
```python
|
| 41 |
+
cd playground/PaintByExample
|
| 42 |
+
python paint_by_example.py
|
| 43 |
+
```
|
| 44 |
+
**Notes:** set `cache_dir` to save the pretrained weights to specific folder. The paint result will be save as `paint_by_example_demo.jpg`:
|
| 45 |
+
|
| 46 |
+
<div align="center">
|
| 47 |
+
|
| 48 |
+
| Input Image | Mask | Example Image | Inpaint Result |
|
| 49 |
+
|:----:|:----:|:----:|:----:|
|
| 50 |
+
|  |  | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/example_image.jpg?raw=true" width=55%></div> |  |
|
| 51 |
+
|  |  | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/pomeranian_example.jpg?raw=true" width=55%></div> |  |
|
| 52 |
+
|
| 53 |
+
</div>
|
| 54 |
+
|
| 55 |
+
### PaintByExample with SAM
|
| 56 |
+
|
| 57 |
+
In this demo, we did inpaint task by:
|
| 58 |
+
1. Generate mask by SAM with prompt (box or point)
|
| 59 |
+
2. Inpaint with mask and example image
|
| 60 |
+
|
| 61 |
+
```python
|
| 62 |
+
cd playground/PaintByExample
|
| 63 |
+
python sam_paint_by_example.py
|
| 64 |
+
```
|
| 65 |
+
**Notes:** We set a more `num_inference_steps` (like 200 to 500) to get higher quality image. And we've found that the mask region can influence a lot on the final result (like a panda can not be well inpainted with a region like dog). It needed to have more test on it.
|
| 66 |
+
|
| 67 |
+
| Input Image | SAM Output | Example Image | Inpaint Result |
|
| 68 |
+
|:----:|:----:|:----:|:----:|
|
| 69 |
+
|  |  | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/labrador_example.jpg?raw=true" width=55%></div> |  |
|
| 70 |
+
|
external/Grounded-Segment-Anything/playground/PaintByExample/paint_by_example.py
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# !pip install diffusers transformers
|
| 2 |
+
|
| 3 |
+
import PIL
|
| 4 |
+
import requests
|
| 5 |
+
import torch
|
| 6 |
+
from io import BytesIO
|
| 7 |
+
from diffusers import DiffusionPipeline
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
"""
|
| 11 |
+
Step 1: Download demo images
|
| 12 |
+
"""
|
| 13 |
+
def download_image(url):
|
| 14 |
+
response = requests.get(url)
|
| 15 |
+
return PIL.Image.open(BytesIO(response.content)).convert("RGB")
|
| 16 |
+
|
| 17 |
+
|
| 18 |
+
img_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/input_image.png?raw=true"
|
| 19 |
+
mask_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/mask.png?raw=true"
|
| 20 |
+
example_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/pomeranian_example.jpg?raw=True"
|
| 21 |
+
# example_url = "https://raw.githubusercontent.com/Fantasy-Studio/Paint-by-Example/main/examples/reference/example_1.jpg"
|
| 22 |
+
|
| 23 |
+
init_image = download_image(img_url).resize((512, 512))
|
| 24 |
+
mask_image = download_image(mask_url).resize((512, 512))
|
| 25 |
+
example_image = download_image(example_url).resize((512, 512))
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
"""
|
| 29 |
+
Step 2: Download pretrained weights and initialize model
|
| 30 |
+
"""
|
| 31 |
+
# set cache dir to store the weights
|
| 32 |
+
cache_dir = "/comp_robot/rentianhe/weights/diffusers/"
|
| 33 |
+
|
| 34 |
+
pipe = DiffusionPipeline.from_pretrained(
|
| 35 |
+
"Fantasy-Studio/Paint-by-Example",
|
| 36 |
+
torch_dtype=torch.float16,
|
| 37 |
+
cache_dir=cache_dir,
|
| 38 |
+
)
|
| 39 |
+
# set to device
|
| 40 |
+
pipe = pipe.to("cuda:1")
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
"""
|
| 44 |
+
Step 3: Run PaintByExample pipeline and save image
|
| 45 |
+
"""
|
| 46 |
+
image = pipe(
|
| 47 |
+
image=init_image,
|
| 48 |
+
mask_image=mask_image,
|
| 49 |
+
example_image=example_image,
|
| 50 |
+
num_inference_steps=200,
|
| 51 |
+
).images[0]
|
| 52 |
+
|
| 53 |
+
image.save("./paint_by_example_demo.jpg")
|
external/Grounded-Segment-Anything/voxelnext_3d_box/models/__init__.py
ADDED
|
File without changes
|
external/Grounded-Segment-Anything/voxelnext_3d_box/models/data_processor.py
ADDED
|
@@ -0,0 +1,211 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from functools import partial
|
| 2 |
+
|
| 3 |
+
import numpy as np
|
| 4 |
+
|
| 5 |
+
tv = None
|
| 6 |
+
try:
|
| 7 |
+
import cumm.tensorview as tv
|
| 8 |
+
except:
|
| 9 |
+
pass
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
def mask_points_by_range(points, limit_range):
|
| 13 |
+
mask = (points[:, 0] >= limit_range[0]) & (points[:, 0] <= limit_range[3]) \
|
| 14 |
+
& (points[:, 1] >= limit_range[1]) & (points[:, 1] <= limit_range[4])
|
| 15 |
+
return mask
|
| 16 |
+
|
| 17 |
+
|
| 18 |
+
class VoxelGeneratorWrapper():
|
| 19 |
+
def __init__(self, vsize_xyz, coors_range_xyz, num_point_features, max_num_points_per_voxel, max_num_voxels):
|
| 20 |
+
try:
|
| 21 |
+
from spconv.utils import VoxelGeneratorV2 as VoxelGenerator
|
| 22 |
+
self.spconv_ver = 1
|
| 23 |
+
except:
|
| 24 |
+
try:
|
| 25 |
+
from spconv.utils import VoxelGenerator
|
| 26 |
+
self.spconv_ver = 1
|
| 27 |
+
except:
|
| 28 |
+
from spconv.utils import Point2VoxelCPU3d as VoxelGenerator
|
| 29 |
+
self.spconv_ver = 2
|
| 30 |
+
|
| 31 |
+
if self.spconv_ver == 1:
|
| 32 |
+
self._voxel_generator = VoxelGenerator(
|
| 33 |
+
voxel_size=vsize_xyz,
|
| 34 |
+
point_cloud_range=coors_range_xyz,
|
| 35 |
+
max_num_points=max_num_points_per_voxel,
|
| 36 |
+
max_voxels=max_num_voxels
|
| 37 |
+
)
|
| 38 |
+
else:
|
| 39 |
+
self._voxel_generator = VoxelGenerator(
|
| 40 |
+
vsize_xyz=vsize_xyz,
|
| 41 |
+
coors_range_xyz=coors_range_xyz,
|
| 42 |
+
num_point_features=num_point_features,
|
| 43 |
+
max_num_points_per_voxel=max_num_points_per_voxel,
|
| 44 |
+
max_num_voxels=max_num_voxels
|
| 45 |
+
)
|
| 46 |
+
|
| 47 |
+
def generate(self, points):
|
| 48 |
+
if self.spconv_ver == 1:
|
| 49 |
+
voxel_output = self._voxel_generator.generate(points)
|
| 50 |
+
if isinstance(voxel_output, dict):
|
| 51 |
+
voxels, coordinates, num_points = \
|
| 52 |
+
voxel_output['voxels'], voxel_output['coordinates'], voxel_output['num_points_per_voxel']
|
| 53 |
+
else:
|
| 54 |
+
voxels, coordinates, num_points = voxel_output
|
| 55 |
+
else:
|
| 56 |
+
assert tv is not None, f"Unexpected error, library: 'cumm' wasn't imported properly."
|
| 57 |
+
voxel_output = self._voxel_generator.point_to_voxel(tv.from_numpy(points))
|
| 58 |
+
tv_voxels, tv_coordinates, tv_num_points = voxel_output
|
| 59 |
+
# make copy with numpy(), since numpy_view() will disappear as soon as the generator is deleted
|
| 60 |
+
voxels = tv_voxels.numpy()
|
| 61 |
+
coordinates = tv_coordinates.numpy()
|
| 62 |
+
num_points = tv_num_points.numpy()
|
| 63 |
+
return voxels, coordinates, num_points
|
| 64 |
+
|
| 65 |
+
|
| 66 |
+
class DataProcessor(object):
|
| 67 |
+
def __init__(self, processor_configs, point_cloud_range, training, num_point_features):
|
| 68 |
+
self.point_cloud_range = point_cloud_range
|
| 69 |
+
self.training = training
|
| 70 |
+
self.num_point_features = num_point_features
|
| 71 |
+
self.mode = 'train' if training else 'test'
|
| 72 |
+
self.grid_size = self.voxel_size = None
|
| 73 |
+
self.data_processor_queue = []
|
| 74 |
+
|
| 75 |
+
self.voxel_generator = None
|
| 76 |
+
|
| 77 |
+
for cur_cfg in processor_configs:
|
| 78 |
+
cur_processor = getattr(self, cur_cfg.NAME)(config=cur_cfg)
|
| 79 |
+
self.data_processor_queue.append(cur_processor)
|
| 80 |
+
|
| 81 |
+
def mask_points_and_boxes_outside_range(self, data_dict=None, config=None):
|
| 82 |
+
if data_dict is None:
|
| 83 |
+
return partial(self.mask_points_and_boxes_outside_range, config=config)
|
| 84 |
+
|
| 85 |
+
if data_dict.get('points', None) is not None:
|
| 86 |
+
mask = mask_points_by_range(data_dict['points'], self.point_cloud_range)
|
| 87 |
+
data_dict['points'] = data_dict['points'][mask]
|
| 88 |
+
|
| 89 |
+
return data_dict
|
| 90 |
+
|
| 91 |
+
def shuffle_points(self, data_dict=None, config=None):
|
| 92 |
+
if data_dict is None:
|
| 93 |
+
return partial(self.shuffle_points, config=config)
|
| 94 |
+
|
| 95 |
+
if config.SHUFFLE_ENABLED[self.mode]:
|
| 96 |
+
points = data_dict['points']
|
| 97 |
+
shuffle_idx = np.random.permutation(points.shape[0])
|
| 98 |
+
points = points[shuffle_idx]
|
| 99 |
+
data_dict['points'] = points
|
| 100 |
+
|
| 101 |
+
return data_dict
|
| 102 |
+
|
| 103 |
+
def transform_points_to_voxels_placeholder(self, data_dict=None, config=None):
|
| 104 |
+
# just calculate grid size
|
| 105 |
+
if data_dict is None:
|
| 106 |
+
grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
|
| 107 |
+
self.grid_size = np.round(grid_size).astype(np.int64)
|
| 108 |
+
self.voxel_size = config.VOXEL_SIZE
|
| 109 |
+
return partial(self.transform_points_to_voxels_placeholder, config=config)
|
| 110 |
+
|
| 111 |
+
return data_dict
|
| 112 |
+
|
| 113 |
+
def double_flip(self, points):
|
| 114 |
+
# y flip
|
| 115 |
+
points_yflip = points.copy()
|
| 116 |
+
points_yflip[:, 1] = -points_yflip[:, 1]
|
| 117 |
+
|
| 118 |
+
# x flip
|
| 119 |
+
points_xflip = points.copy()
|
| 120 |
+
points_xflip[:, 0] = -points_xflip[:, 0]
|
| 121 |
+
|
| 122 |
+
# x y flip
|
| 123 |
+
points_xyflip = points.copy()
|
| 124 |
+
points_xyflip[:, 0] = -points_xyflip[:, 0]
|
| 125 |
+
points_xyflip[:, 1] = -points_xyflip[:, 1]
|
| 126 |
+
|
| 127 |
+
return points_yflip, points_xflip, points_xyflip
|
| 128 |
+
|
| 129 |
+
def transform_points_to_voxels(self, data_dict=None, config=None):
|
| 130 |
+
if data_dict is None:
|
| 131 |
+
grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
|
| 132 |
+
self.grid_size = np.round(grid_size).astype(np.int64)
|
| 133 |
+
self.voxel_size = config.VOXEL_SIZE
|
| 134 |
+
# just bind the config, we will create the VoxelGeneratorWrapper later,
|
| 135 |
+
# to avoid pickling issues in multiprocess spawn
|
| 136 |
+
return partial(self.transform_points_to_voxels, config=config)
|
| 137 |
+
|
| 138 |
+
if self.voxel_generator is None:
|
| 139 |
+
self.voxel_generator = VoxelGeneratorWrapper(
|
| 140 |
+
vsize_xyz=config.VOXEL_SIZE,
|
| 141 |
+
coors_range_xyz=self.point_cloud_range,
|
| 142 |
+
num_point_features=self.num_point_features,
|
| 143 |
+
max_num_points_per_voxel=config.MAX_POINTS_PER_VOXEL,
|
| 144 |
+
max_num_voxels=config.MAX_NUMBER_OF_VOXELS[self.mode],
|
| 145 |
+
)
|
| 146 |
+
|
| 147 |
+
points = data_dict['points']
|
| 148 |
+
voxel_output = self.voxel_generator.generate(points)
|
| 149 |
+
voxels, coordinates, num_points = voxel_output
|
| 150 |
+
|
| 151 |
+
data_dict['voxels'] = voxels
|
| 152 |
+
data_dict['voxel_coords'] = coordinates
|
| 153 |
+
data_dict['voxel_num_points'] = num_points
|
| 154 |
+
return data_dict
|
| 155 |
+
|
| 156 |
+
def sample_points(self, data_dict=None, config=None):
|
| 157 |
+
if data_dict is None:
|
| 158 |
+
return partial(self.sample_points, config=config)
|
| 159 |
+
|
| 160 |
+
num_points = config.NUM_POINTS[self.mode]
|
| 161 |
+
if num_points == -1:
|
| 162 |
+
return data_dict
|
| 163 |
+
|
| 164 |
+
points = data_dict['points']
|
| 165 |
+
if num_points < len(points):
|
| 166 |
+
pts_depth = np.linalg.norm(points[:, 0:3], axis=1)
|
| 167 |
+
pts_near_flag = pts_depth < 40.0
|
| 168 |
+
far_idxs_choice = np.where(pts_near_flag == 0)[0]
|
| 169 |
+
near_idxs = np.where(pts_near_flag == 1)[0]
|
| 170 |
+
choice = []
|
| 171 |
+
if num_points > len(far_idxs_choice):
|
| 172 |
+
near_idxs_choice = np.random.choice(near_idxs, num_points - len(far_idxs_choice), replace=False)
|
| 173 |
+
choice = np.concatenate((near_idxs_choice, far_idxs_choice), axis=0) \
|
| 174 |
+
if len(far_idxs_choice) > 0 else near_idxs_choice
|
| 175 |
+
else:
|
| 176 |
+
choice = np.arange(0, len(points), dtype=np.int32)
|
| 177 |
+
choice = np.random.choice(choice, num_points, replace=False)
|
| 178 |
+
np.random.shuffle(choice)
|
| 179 |
+
else:
|
| 180 |
+
choice = np.arange(0, len(points), dtype=np.int32)
|
| 181 |
+
if num_points > len(points):
|
| 182 |
+
extra_choice = np.random.choice(choice, num_points - len(points), replace=False)
|
| 183 |
+
choice = np.concatenate((choice, extra_choice), axis=0)
|
| 184 |
+
np.random.shuffle(choice)
|
| 185 |
+
data_dict['points'] = points[choice]
|
| 186 |
+
return data_dict
|
| 187 |
+
|
| 188 |
+
def calculate_grid_size(self, data_dict=None, config=None):
|
| 189 |
+
if data_dict is None:
|
| 190 |
+
grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
|
| 191 |
+
self.grid_size = np.round(grid_size).astype(np.int64)
|
| 192 |
+
self.voxel_size = config.VOXEL_SIZE
|
| 193 |
+
return partial(self.calculate_grid_size, config=config)
|
| 194 |
+
return data_dict
|
| 195 |
+
|
| 196 |
+
def forward(self, data_dict):
|
| 197 |
+
"""
|
| 198 |
+
Args:
|
| 199 |
+
data_dict:
|
| 200 |
+
points: (N, 3 + C_in)
|
| 201 |
+
gt_boxes: optional, (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...]
|
| 202 |
+
gt_names: optional, (N), string
|
| 203 |
+
...
|
| 204 |
+
|
| 205 |
+
Returns:
|
| 206 |
+
"""
|
| 207 |
+
|
| 208 |
+
for cur_processor in self.data_processor_queue:
|
| 209 |
+
data_dict = cur_processor(data_dict=data_dict)
|
| 210 |
+
|
| 211 |
+
return data_dict
|
external/Grounded-Segment-Anything/voxelnext_3d_box/models/mean_vfe.py
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
class MeanVFE(nn.Module):
|
| 5 |
+
def __init__(self):
|
| 6 |
+
super().__init__()
|
| 7 |
+
|
| 8 |
+
def forward(self, batch_dict, **kwargs):
|
| 9 |
+
"""
|
| 10 |
+
Args:
|
| 11 |
+
batch_dict:
|
| 12 |
+
voxels: (num_voxels, max_points_per_voxel, C)
|
| 13 |
+
voxel_num_points: optional (num_voxels)
|
| 14 |
+
**kwargs:
|
| 15 |
+
|
| 16 |
+
Returns:
|
| 17 |
+
vfe_features: (num_voxels, C)
|
| 18 |
+
"""
|
| 19 |
+
voxel_features, voxel_num_points = batch_dict['voxels'], batch_dict['voxel_num_points']
|
| 20 |
+
points_mean = voxel_features[:, :, :].sum(dim=1, keepdim=False)
|
| 21 |
+
normalizer = torch.clamp_min(voxel_num_points.view(-1, 1), min=1.0).type_as(voxel_features)
|
| 22 |
+
points_mean = points_mean / normalizer
|
| 23 |
+
batch_dict['voxel_features'] = points_mean.contiguous()
|
| 24 |
+
|
| 25 |
+
return batch_dict
|
external/Grounded-Segment-Anything/voxelnext_3d_box/models/spconv_backbone_voxelnext.py
ADDED
|
@@ -0,0 +1,292 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from functools import partial
|
| 2 |
+
import torch
|
| 3 |
+
import torch.nn as nn
|
| 4 |
+
|
| 5 |
+
import spconv.pytorch as spconv
|
| 6 |
+
from spconv.core import ConvAlgo
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
def replace_feature(out, new_features):
|
| 10 |
+
return out.replace_feature(new_features)
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
def post_act_block(in_channels, out_channels, kernel_size, indice_key=None, stride=1, padding=0,
|
| 14 |
+
conv_type='subm', norm_fn=None):
|
| 15 |
+
|
| 16 |
+
if conv_type == 'subm':
|
| 17 |
+
conv = spconv.SubMConv3d(in_channels, out_channels, kernel_size, bias=False, indice_key=indice_key, algo=ConvAlgo.Native)
|
| 18 |
+
elif conv_type == 'spconv':
|
| 19 |
+
conv = spconv.SparseConv3d(in_channels, out_channels, kernel_size, stride=stride, padding=padding,
|
| 20 |
+
bias=False, indice_key=indice_key, algo=ConvAlgo.Native)
|
| 21 |
+
elif conv_type == 'inverseconv':
|
| 22 |
+
conv = spconv.SparseInverseConv3d(in_channels, out_channels, kernel_size, indice_key=indice_key, bias=False, algo=ConvAlgo.Native)
|
| 23 |
+
else:
|
| 24 |
+
raise NotImplementedError
|
| 25 |
+
|
| 26 |
+
m = spconv.SparseSequential(
|
| 27 |
+
conv,
|
| 28 |
+
norm_fn(out_channels),
|
| 29 |
+
nn.ReLU(),
|
| 30 |
+
)
|
| 31 |
+
|
| 32 |
+
return m
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
class SparseBasicBlock(spconv.SparseModule):
|
| 36 |
+
expansion = 1
|
| 37 |
+
|
| 38 |
+
def __init__(self, inplanes, planes, stride=1, norm_fn=None, downsample=None, indice_key=None):
|
| 39 |
+
super(SparseBasicBlock, self).__init__()
|
| 40 |
+
|
| 41 |
+
assert norm_fn is not None
|
| 42 |
+
bias = norm_fn is not None
|
| 43 |
+
self.conv1 = spconv.SubMConv3d(
|
| 44 |
+
inplanes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key, algo=ConvAlgo.Native
|
| 45 |
+
)
|
| 46 |
+
self.bn1 = norm_fn(planes)
|
| 47 |
+
self.relu = nn.ReLU()
|
| 48 |
+
self.conv2 = spconv.SubMConv3d(
|
| 49 |
+
planes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key, algo=ConvAlgo.Native
|
| 50 |
+
)
|
| 51 |
+
self.bn2 = norm_fn(planes)
|
| 52 |
+
self.downsample = downsample
|
| 53 |
+
self.stride = stride
|
| 54 |
+
|
| 55 |
+
def forward(self, x):
|
| 56 |
+
identity = x
|
| 57 |
+
|
| 58 |
+
out = self.conv1(x)
|
| 59 |
+
out = replace_feature(out, self.bn1(out.features))
|
| 60 |
+
out = replace_feature(out, self.relu(out.features))
|
| 61 |
+
|
| 62 |
+
out = self.conv2(out)
|
| 63 |
+
out = replace_feature(out, self.bn2(out.features))
|
| 64 |
+
|
| 65 |
+
if self.downsample is not None:
|
| 66 |
+
identity = self.downsample(x)
|
| 67 |
+
|
| 68 |
+
out = replace_feature(out, out.features + identity.features)
|
| 69 |
+
out = replace_feature(out, self.relu(out.features))
|
| 70 |
+
|
| 71 |
+
return out
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
class VoxelResBackBone8xVoxelNeXt(nn.Module):
|
| 75 |
+
def __init__(self, input_channels, grid_size, **kwargs):
|
| 76 |
+
super().__init__()
|
| 77 |
+
norm_fn = partial(nn.BatchNorm1d, eps=1e-3, momentum=0.01)
|
| 78 |
+
|
| 79 |
+
spconv_kernel_sizes = [3, 3, 3, 3]
|
| 80 |
+
channels = [16, 32, 64, 128, 128]
|
| 81 |
+
out_channel = 128
|
| 82 |
+
|
| 83 |
+
self.sparse_shape = grid_size[::-1] + [1, 0, 0]
|
| 84 |
+
|
| 85 |
+
self.conv_input = spconv.SparseSequential(
|
| 86 |
+
spconv.SubMConv3d(input_channels, channels[0], 3, padding=1, bias=False, indice_key='subm1', algo=ConvAlgo.Native),
|
| 87 |
+
norm_fn(channels[0]),
|
| 88 |
+
nn.ReLU(),
|
| 89 |
+
)
|
| 90 |
+
block = post_act_block
|
| 91 |
+
|
| 92 |
+
self.conv1 = spconv.SparseSequential(
|
| 93 |
+
SparseBasicBlock(channels[0], channels[0], norm_fn=norm_fn, indice_key='res1'),
|
| 94 |
+
SparseBasicBlock(channels[0], channels[0], norm_fn=norm_fn, indice_key='res1'),
|
| 95 |
+
)
|
| 96 |
+
|
| 97 |
+
self.conv2 = spconv.SparseSequential(
|
| 98 |
+
# [1600, 1408, 41] <- [800, 704, 21]
|
| 99 |
+
block(channels[0], channels[1], spconv_kernel_sizes[0], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[0]//2), indice_key='spconv2', conv_type='spconv'),
|
| 100 |
+
SparseBasicBlock(channels[1], channels[1], norm_fn=norm_fn, indice_key='res2'),
|
| 101 |
+
SparseBasicBlock(channels[1], channels[1], norm_fn=norm_fn, indice_key='res2'),
|
| 102 |
+
)
|
| 103 |
+
|
| 104 |
+
self.conv3 = spconv.SparseSequential(
|
| 105 |
+
# [800, 704, 21] <- [400, 352, 11]
|
| 106 |
+
block(channels[1], channels[2], spconv_kernel_sizes[1], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[1]//2), indice_key='spconv3', conv_type='spconv'),
|
| 107 |
+
SparseBasicBlock(channels[2], channels[2], norm_fn=norm_fn, indice_key='res3'),
|
| 108 |
+
SparseBasicBlock(channels[2], channels[2], norm_fn=norm_fn, indice_key='res3'),
|
| 109 |
+
)
|
| 110 |
+
|
| 111 |
+
self.conv4 = spconv.SparseSequential(
|
| 112 |
+
# [400, 352, 11] <- [200, 176, 6]
|
| 113 |
+
block(channels[2], channels[3], spconv_kernel_sizes[2], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[2]//2), indice_key='spconv4', conv_type='spconv'),
|
| 114 |
+
SparseBasicBlock(channels[3], channels[3], norm_fn=norm_fn, indice_key='res4'),
|
| 115 |
+
SparseBasicBlock(channels[3], channels[3], norm_fn=norm_fn, indice_key='res4'),
|
| 116 |
+
)
|
| 117 |
+
|
| 118 |
+
self.conv5 = spconv.SparseSequential(
|
| 119 |
+
# [200, 176, 6] <- [100, 88, 3]
|
| 120 |
+
block(channels[3], channels[4], spconv_kernel_sizes[3], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[3]//2), indice_key='spconv5', conv_type='spconv'),
|
| 121 |
+
SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res5'),
|
| 122 |
+
SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res5'),
|
| 123 |
+
)
|
| 124 |
+
|
| 125 |
+
self.conv6 = spconv.SparseSequential(
|
| 126 |
+
# [200, 176, 6] <- [100, 88, 3]
|
| 127 |
+
block(channels[4], channels[4], spconv_kernel_sizes[3], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[3]//2), indice_key='spconv6', conv_type='spconv'),
|
| 128 |
+
SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res6'),
|
| 129 |
+
SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res6'),
|
| 130 |
+
)
|
| 131 |
+
self.conv_out = spconv.SparseSequential(
|
| 132 |
+
# [200, 150, 5] -> [200, 150, 2]
|
| 133 |
+
spconv.SparseConv2d(channels[3], out_channel, 3, stride=1, padding=1, bias=False, indice_key='spconv_down2', algo=ConvAlgo.Native),
|
| 134 |
+
norm_fn(out_channel),
|
| 135 |
+
nn.ReLU(),
|
| 136 |
+
)
|
| 137 |
+
|
| 138 |
+
self.shared_conv = spconv.SparseSequential(
|
| 139 |
+
spconv.SubMConv2d(out_channel, out_channel, 3, stride=1, padding=1, bias=True, algo=ConvAlgo.Native),
|
| 140 |
+
nn.BatchNorm1d(out_channel),
|
| 141 |
+
nn.ReLU(True),
|
| 142 |
+
)
|
| 143 |
+
|
| 144 |
+
self.forward_ret_dict = {}
|
| 145 |
+
self.num_point_features = out_channel
|
| 146 |
+
self.backbone_channels = {
|
| 147 |
+
'x_conv1': channels[0],
|
| 148 |
+
'x_conv2': channels[1],
|
| 149 |
+
'x_conv3': channels[2],
|
| 150 |
+
'x_conv4': channels[3]
|
| 151 |
+
}
|
| 152 |
+
|
| 153 |
+
def bev_out(self, x_conv, index):
|
| 154 |
+
features_cat = x_conv.features
|
| 155 |
+
indices_cat = x_conv.indices[:, [0, 2, 3]]
|
| 156 |
+
spatial_shape = x_conv.spatial_shape[1:]
|
| 157 |
+
|
| 158 |
+
indices_unique, _inv = torch.unique(indices_cat, dim=0, return_inverse=True)
|
| 159 |
+
features_unique = features_cat.new_zeros((indices_unique.shape[0], features_cat.shape[1]))
|
| 160 |
+
features_unique.index_add_(0, _inv, features_cat)
|
| 161 |
+
|
| 162 |
+
perm = torch.arange(_inv.size(0), dtype=_inv.dtype, device=_inv.device)
|
| 163 |
+
perm = _inv.new_empty(indices_unique.size(0)).scatter_(0, _inv, perm)
|
| 164 |
+
index_out = index[perm]
|
| 165 |
+
|
| 166 |
+
x_out = spconv.SparseConvTensor(
|
| 167 |
+
features=features_unique,
|
| 168 |
+
indices=indices_unique,
|
| 169 |
+
spatial_shape=spatial_shape,
|
| 170 |
+
batch_size=x_conv.batch_size
|
| 171 |
+
)
|
| 172 |
+
return x_out, index_out
|
| 173 |
+
|
| 174 |
+
def track_voxels_2d(self, x, x_downsample, index, kernel_size=3):
|
| 175 |
+
_step = int(kernel_size//2)
|
| 176 |
+
kernel_offsets = [[i, j] for i in range(-_step, _step+1) for j in range(-_step, _step+1)]
|
| 177 |
+
#kernel_offsets.remove([0, 0])
|
| 178 |
+
kernel_offsets = torch.Tensor(kernel_offsets).to(x.indices.device)
|
| 179 |
+
|
| 180 |
+
batch_size = x.batch_size
|
| 181 |
+
index_batch = []
|
| 182 |
+
indices_batch = []
|
| 183 |
+
|
| 184 |
+
for b in range(batch_size):
|
| 185 |
+
batch_index = x.indices[:, 0]==b
|
| 186 |
+
indices_ori = x.indices[batch_index]
|
| 187 |
+
features_ori = index[batch_index]
|
| 188 |
+
|
| 189 |
+
features_fore = features_ori
|
| 190 |
+
coords_fore = indices_ori
|
| 191 |
+
|
| 192 |
+
voxel_kerels_imp = kernel_offsets.unsqueeze(0).repeat(features_fore.shape[0],1, 1)
|
| 193 |
+
indices_fore_kernels = coords_fore[:, 1:].unsqueeze(1).repeat(1, kernel_offsets.shape[0], 1)
|
| 194 |
+
indices_with_imp = indices_fore_kernels + voxel_kerels_imp
|
| 195 |
+
features_fore = features_fore.repeat(1, kernel_offsets.shape[0])
|
| 196 |
+
|
| 197 |
+
selected_indices = indices_with_imp
|
| 198 |
+
spatial_indices = (selected_indices[:, :, 0] >=0) * (selected_indices[:, :, 1] >=0) * \
|
| 199 |
+
(selected_indices[:, :, 0] < x.spatial_shape[0]) * (selected_indices[:, :, 1] < x.spatial_shape[1])
|
| 200 |
+
selected_indices = selected_indices[spatial_indices]
|
| 201 |
+
features_fore = features_fore[spatial_indices].view(-1, 1)
|
| 202 |
+
|
| 203 |
+
selected_indices = torch.cat([torch.ones((selected_indices.shape[0], 1), device=features_fore.device)*b, selected_indices], dim=1)
|
| 204 |
+
|
| 205 |
+
features_fore, coords_fore = features_fore, selected_indices
|
| 206 |
+
index_batch.append(features_fore)
|
| 207 |
+
indices_batch.append(coords_fore)
|
| 208 |
+
|
| 209 |
+
index_batch = torch.cat(index_batch)
|
| 210 |
+
indices_batch = torch.cat(indices_batch)
|
| 211 |
+
|
| 212 |
+
return self.index_from_sparse(index_batch, indices_batch, x_downsample, True)
|
| 213 |
+
|
| 214 |
+
def index_from_sparse(self, feature, indices, x_target, _2d=False):
|
| 215 |
+
sparse_index = spconv.SparseConvTensor(
|
| 216 |
+
features=feature,
|
| 217 |
+
indices=indices.int(),
|
| 218 |
+
spatial_shape=x_target.spatial_shape,
|
| 219 |
+
batch_size=x_target.batch_size
|
| 220 |
+
)
|
| 221 |
+
dense_index = sparse_index.dense()
|
| 222 |
+
indices_downsample = x_target.indices.long()
|
| 223 |
+
if _2d:
|
| 224 |
+
index_downsample = dense_index[indices_downsample[:, 0], :, indices_downsample[:, 1], indices_downsample[:, 2]]
|
| 225 |
+
else:
|
| 226 |
+
index_downsample = dense_index[indices_downsample[:, 0], :, indices_downsample[:, 1], indices_downsample[:, 2], indices_downsample[:, 3]]
|
| 227 |
+
return index_downsample
|
| 228 |
+
|
| 229 |
+
def forward(self, batch_dict):
|
| 230 |
+
"""
|
| 231 |
+
Args:
|
| 232 |
+
batch_dict:
|
| 233 |
+
batch_size: int
|
| 234 |
+
vfe_features: (num_voxels, C)
|
| 235 |
+
voxel_coords: (num_voxels, 4), [batch_idx, z_idx, y_idx, x_idx]
|
| 236 |
+
Returns:
|
| 237 |
+
batch_dict:
|
| 238 |
+
encoded_spconv_tensor: sparse tensor
|
| 239 |
+
"""
|
| 240 |
+
voxel_features, voxel_coords = batch_dict['voxel_features'], batch_dict['voxel_coords']
|
| 241 |
+
batch_size = batch_dict['batch_size']
|
| 242 |
+
input_sp_tensor = spconv.SparseConvTensor(
|
| 243 |
+
features=voxel_features,
|
| 244 |
+
indices=voxel_coords.int(),
|
| 245 |
+
spatial_shape=self.sparse_shape,
|
| 246 |
+
batch_size=batch_size
|
| 247 |
+
)
|
| 248 |
+
x = self.conv_input(input_sp_tensor)
|
| 249 |
+
|
| 250 |
+
x_conv1 = self.conv1(x)
|
| 251 |
+
x_conv2 = self.conv2(x_conv1)
|
| 252 |
+
x_conv3 = self.conv3(x_conv2)
|
| 253 |
+
x_conv4 = self.conv4(x_conv3)
|
| 254 |
+
x_conv5 = self.conv5(x_conv4)
|
| 255 |
+
x_conv6 = self.conv6(x_conv5)
|
| 256 |
+
|
| 257 |
+
x_conv5.indices[:, 1:] *= 2
|
| 258 |
+
x_conv6.indices[:, 1:] *= 4
|
| 259 |
+
x_conv4 = x_conv4.replace_feature(torch.cat([x_conv4.features, x_conv5.features, x_conv6.features]))
|
| 260 |
+
x_conv4.indices = torch.cat([x_conv4.indices, x_conv5.indices, x_conv6.indices])
|
| 261 |
+
|
| 262 |
+
index6_out = torch.arange(x_conv4.indices.shape[0], device=x_conv4.indices.device).unsqueeze(-1)
|
| 263 |
+
out_bevout, index_bevout = self.bev_out(x_conv4, index6_out)
|
| 264 |
+
|
| 265 |
+
out = self.conv_out(out_bevout)
|
| 266 |
+
index_out = self.track_voxels_2d(out_bevout, out, index_bevout)
|
| 267 |
+
|
| 268 |
+
out = self.shared_conv(out)
|
| 269 |
+
|
| 270 |
+
batch_dict.update({
|
| 271 |
+
'encoded_spconv_tensor': out,
|
| 272 |
+
'encoded_spconv_tensor_stride': 8,
|
| 273 |
+
'out_voxels': x_conv4.indices[index_out.squeeze(-1)],
|
| 274 |
+
})
|
| 275 |
+
batch_dict.update({
|
| 276 |
+
'multi_scale_3d_features': {
|
| 277 |
+
'x_conv1': x_conv1,
|
| 278 |
+
'x_conv2': x_conv2,
|
| 279 |
+
'x_conv3': x_conv3,
|
| 280 |
+
'x_conv4': x_conv4,
|
| 281 |
+
}
|
| 282 |
+
})
|
| 283 |
+
batch_dict.update({
|
| 284 |
+
'multi_scale_3d_strides': {
|
| 285 |
+
'x_conv1': 1,
|
| 286 |
+
'x_conv2': 2,
|
| 287 |
+
'x_conv3': 4,
|
| 288 |
+
'x_conv4': 8,
|
| 289 |
+
}
|
| 290 |
+
})
|
| 291 |
+
|
| 292 |
+
return batch_dict
|
external/Grounded-Segment-Anything/voxelnext_3d_box/models/voxelnext_head.py
ADDED
|
@@ -0,0 +1,166 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import numpy as np
|
| 2 |
+
import torch
|
| 3 |
+
import torch.nn as nn
|
| 4 |
+
from voxelnext_3d_box.utils import centernet_utils
|
| 5 |
+
import spconv.pytorch as spconv
|
| 6 |
+
import copy
|
| 7 |
+
from spconv.core import ConvAlgo
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
class SeparateHead(nn.Module):
|
| 11 |
+
def __init__(self, input_channels, sep_head_dict, kernel_size, use_bias=False):
|
| 12 |
+
super().__init__()
|
| 13 |
+
self.sep_head_dict = sep_head_dict
|
| 14 |
+
|
| 15 |
+
for cur_name in self.sep_head_dict:
|
| 16 |
+
output_channels = self.sep_head_dict[cur_name]['out_channels']
|
| 17 |
+
num_conv = self.sep_head_dict[cur_name]['num_conv']
|
| 18 |
+
|
| 19 |
+
fc_list = []
|
| 20 |
+
for k in range(num_conv - 1):
|
| 21 |
+
fc_list.append(spconv.SparseSequential(
|
| 22 |
+
spconv.SubMConv2d(input_channels, input_channels, kernel_size, padding=int(kernel_size//2), bias=use_bias, indice_key=cur_name, algo=ConvAlgo.Native),
|
| 23 |
+
nn.BatchNorm1d(input_channels),
|
| 24 |
+
nn.ReLU()
|
| 25 |
+
))
|
| 26 |
+
fc_list.append(spconv.SubMConv2d(input_channels, output_channels, 1, bias=True, indice_key=cur_name+'out', algo=ConvAlgo.Native))
|
| 27 |
+
fc = nn.Sequential(*fc_list)
|
| 28 |
+
self.__setattr__(cur_name, fc)
|
| 29 |
+
|
| 30 |
+
def forward(self, x):
|
| 31 |
+
ret_dict = {}
|
| 32 |
+
for cur_name in self.sep_head_dict:
|
| 33 |
+
ret_dict[cur_name] = self.__getattr__(cur_name)(x).features
|
| 34 |
+
|
| 35 |
+
return ret_dict
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
class VoxelNeXtHead(nn.Module):
|
| 39 |
+
def __init__(self, class_names, point_cloud_range, voxel_size, kernel_size_head,
|
| 40 |
+
CLASS_NAMES_EACH_HEAD, SEPARATE_HEAD_CFG, POST_PROCESSING):
|
| 41 |
+
super().__init__()
|
| 42 |
+
self.point_cloud_range = torch.Tensor(point_cloud_range)
|
| 43 |
+
self.voxel_size = torch.Tensor(voxel_size)
|
| 44 |
+
self.feature_map_stride = 8
|
| 45 |
+
|
| 46 |
+
self.class_names = class_names
|
| 47 |
+
self.class_names_each_head = []
|
| 48 |
+
self.class_id_mapping_each_head = []
|
| 49 |
+
self.POST_PROCESSING = POST_PROCESSING
|
| 50 |
+
|
| 51 |
+
for cur_class_names in CLASS_NAMES_EACH_HEAD:
|
| 52 |
+
self.class_names_each_head.append([x for x in cur_class_names if x in class_names])
|
| 53 |
+
cur_class_id_mapping = torch.from_numpy(np.array(
|
| 54 |
+
[self.class_names.index(x) for x in cur_class_names if x in class_names]
|
| 55 |
+
))
|
| 56 |
+
self.class_id_mapping_each_head.append(cur_class_id_mapping)
|
| 57 |
+
|
| 58 |
+
total_classes = sum([len(x) for x in self.class_names_each_head])
|
| 59 |
+
assert total_classes == len(self.class_names), f'class_names_each_head={self.class_names_each_head}'
|
| 60 |
+
|
| 61 |
+
self.heads_list = nn.ModuleList()
|
| 62 |
+
self.separate_head_cfg = SEPARATE_HEAD_CFG
|
| 63 |
+
for idx, cur_class_names in enumerate(self.class_names_each_head):
|
| 64 |
+
cur_head_dict = copy.deepcopy(self.separate_head_cfg.HEAD_DICT)
|
| 65 |
+
cur_head_dict['hm'] = dict(out_channels=len(cur_class_names), num_conv=2)
|
| 66 |
+
self.heads_list.append(
|
| 67 |
+
SeparateHead(
|
| 68 |
+
input_channels=128,
|
| 69 |
+
sep_head_dict=cur_head_dict,
|
| 70 |
+
kernel_size=kernel_size_head,
|
| 71 |
+
use_bias=True,
|
| 72 |
+
)
|
| 73 |
+
)
|
| 74 |
+
self.forward_ret_dict = {}
|
| 75 |
+
|
| 76 |
+
def generate_predicted_boxes(self, batch_size, pred_dicts, voxel_indices, spatial_shape):
|
| 77 |
+
device = pred_dicts[0]['hm'].device
|
| 78 |
+
post_process_cfg = self.POST_PROCESSING
|
| 79 |
+
post_center_limit_range = torch.tensor(post_process_cfg.POST_CENTER_LIMIT_RANGE).float().to(device)
|
| 80 |
+
|
| 81 |
+
ret_dict = [{
|
| 82 |
+
'pred_boxes': [],
|
| 83 |
+
'pred_scores': [],
|
| 84 |
+
'pred_labels': [],
|
| 85 |
+
'pred_ious': [],
|
| 86 |
+
'voxel_ids': []
|
| 87 |
+
} for k in range(batch_size)]
|
| 88 |
+
for idx, pred_dict in enumerate(pred_dicts):
|
| 89 |
+
batch_hm = pred_dict['hm'].sigmoid()
|
| 90 |
+
batch_center = pred_dict['center']
|
| 91 |
+
batch_center_z = pred_dict['center_z']
|
| 92 |
+
batch_dim = pred_dict['dim'].exp()
|
| 93 |
+
batch_rot_cos = pred_dict['rot'][:, 0].unsqueeze(dim=1)
|
| 94 |
+
batch_rot_sin = pred_dict['rot'][:, 1].unsqueeze(dim=1)
|
| 95 |
+
batch_iou = None
|
| 96 |
+
batch_vel = pred_dict['vel'] if 'vel' in self.separate_head_cfg.HEAD_ORDER else None
|
| 97 |
+
voxel_indices_ = voxel_indices
|
| 98 |
+
|
| 99 |
+
final_pred_dicts = centernet_utils.decode_bbox_from_voxels_nuscenes(
|
| 100 |
+
batch_size=batch_size, indices=voxel_indices_,
|
| 101 |
+
obj=batch_hm,
|
| 102 |
+
rot_cos=batch_rot_cos,
|
| 103 |
+
rot_sin=batch_rot_sin,
|
| 104 |
+
center=batch_center, center_z=batch_center_z,
|
| 105 |
+
dim=batch_dim, vel=batch_vel, iou=batch_iou,
|
| 106 |
+
point_cloud_range=self.point_cloud_range.to(device), voxel_size=self.voxel_size.to(device),
|
| 107 |
+
feature_map_stride=self.feature_map_stride,
|
| 108 |
+
K=post_process_cfg.MAX_OBJ_PER_SAMPLE,
|
| 109 |
+
score_thresh=post_process_cfg.SCORE_THRESH,
|
| 110 |
+
post_center_limit_range=post_center_limit_range,
|
| 111 |
+
add_features=torch.arange(voxel_indices_.shape[0], device=voxel_indices_.device).unsqueeze(-1)
|
| 112 |
+
)
|
| 113 |
+
|
| 114 |
+
for k, final_dict in enumerate(final_pred_dicts):
|
| 115 |
+
class_id_mapping_each_head = self.class_id_mapping_each_head[idx].to(device)
|
| 116 |
+
final_dict['pred_labels'] = class_id_mapping_each_head[final_dict['pred_labels'].long()]
|
| 117 |
+
|
| 118 |
+
ret_dict[k]['pred_boxes'].append(final_dict['pred_boxes'])
|
| 119 |
+
ret_dict[k]['pred_scores'].append(final_dict['pred_scores'])
|
| 120 |
+
ret_dict[k]['pred_labels'].append(final_dict['pred_labels'])
|
| 121 |
+
ret_dict[k]['pred_ious'].append(final_dict['pred_ious'])
|
| 122 |
+
ret_dict[k]['voxel_ids'].append(final_dict['add_features'])
|
| 123 |
+
|
| 124 |
+
for k in range(batch_size):
|
| 125 |
+
pred_boxes = torch.cat(ret_dict[k]['pred_boxes'], dim=0)
|
| 126 |
+
pred_scores = torch.cat(ret_dict[k]['pred_scores'], dim=0)
|
| 127 |
+
pred_labels = torch.cat(ret_dict[k]['pred_labels'], dim=0)
|
| 128 |
+
voxel_ids = torch.cat(ret_dict[k]['voxel_ids'], dim=0)
|
| 129 |
+
|
| 130 |
+
ret_dict[k]['pred_boxes'] = pred_boxes
|
| 131 |
+
ret_dict[k]['pred_scores'] = pred_scores
|
| 132 |
+
ret_dict[k]['pred_labels'] = pred_labels + 1
|
| 133 |
+
ret_dict[k]['voxel_ids'] = voxel_ids
|
| 134 |
+
|
| 135 |
+
return ret_dict
|
| 136 |
+
|
| 137 |
+
def _get_voxel_infos(self, x):
|
| 138 |
+
spatial_shape = x.spatial_shape
|
| 139 |
+
voxel_indices = x.indices
|
| 140 |
+
spatial_indices = []
|
| 141 |
+
num_voxels = []
|
| 142 |
+
batch_size = x.batch_size
|
| 143 |
+
batch_index = voxel_indices[:, 0]
|
| 144 |
+
|
| 145 |
+
for bs_idx in range(batch_size):
|
| 146 |
+
batch_inds = batch_index==bs_idx
|
| 147 |
+
spatial_indices.append(voxel_indices[batch_inds][:, [2, 1]])
|
| 148 |
+
num_voxels.append(batch_inds.sum())
|
| 149 |
+
|
| 150 |
+
return spatial_shape, batch_index, voxel_indices, spatial_indices, num_voxels
|
| 151 |
+
|
| 152 |
+
def forward(self, data_dict):
|
| 153 |
+
x = data_dict['encoded_spconv_tensor']
|
| 154 |
+
spatial_shape, batch_index, voxel_indices, spatial_indices, num_voxels = self._get_voxel_infos(x)
|
| 155 |
+
|
| 156 |
+
pred_dicts = []
|
| 157 |
+
for idx, head in enumerate(self.heads_list):
|
| 158 |
+
pred_dict = head(x)
|
| 159 |
+
pred_dicts.append(pred_dict)
|
| 160 |
+
|
| 161 |
+
pred_dicts = self.generate_predicted_boxes(
|
| 162 |
+
data_dict['batch_size'],
|
| 163 |
+
pred_dicts, voxel_indices, spatial_shape
|
| 164 |
+
)
|
| 165 |
+
|
| 166 |
+
return pred_dicts
|
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/centernet_utils.py
ADDED
|
@@ -0,0 +1,116 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is modified from https://github.com/tianweiy/CenterPoint
|
| 2 |
+
|
| 3 |
+
import torch
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
def _topk_1d(scores, batch_size, batch_idx, obj, K=40, nuscenes=False):
|
| 7 |
+
# scores: (N, num_classes)
|
| 8 |
+
topk_score_list = []
|
| 9 |
+
topk_inds_list = []
|
| 10 |
+
topk_classes_list = []
|
| 11 |
+
|
| 12 |
+
for bs_idx in range(batch_size):
|
| 13 |
+
batch_inds = batch_idx==bs_idx
|
| 14 |
+
if obj.shape[-1] == 1 and not nuscenes:
|
| 15 |
+
score = scores[batch_inds].permute(1, 0)
|
| 16 |
+
topk_scores, topk_inds = torch.topk(score, K)
|
| 17 |
+
topk_score, topk_ind = torch.topk(obj[topk_inds.view(-1)].squeeze(-1), K) #torch.topk(topk_scores.view(-1), K)
|
| 18 |
+
else:
|
| 19 |
+
score = obj[batch_inds].permute(1, 0)
|
| 20 |
+
topk_scores, topk_inds = torch.topk(score, min(K, score.shape[-1]))
|
| 21 |
+
topk_score, topk_ind = torch.topk(topk_scores.view(-1), min(K, topk_scores.view(-1).shape[-1]))
|
| 22 |
+
|
| 23 |
+
topk_classes = (topk_ind // K).int()
|
| 24 |
+
topk_inds = topk_inds.view(-1).gather(0, topk_ind)
|
| 25 |
+
#print('topk_inds', topk_inds)
|
| 26 |
+
|
| 27 |
+
if not obj is None and obj.shape[-1] == 1:
|
| 28 |
+
topk_score_list.append(obj[batch_inds][topk_inds])
|
| 29 |
+
else:
|
| 30 |
+
topk_score_list.append(topk_score)
|
| 31 |
+
topk_inds_list.append(topk_inds)
|
| 32 |
+
topk_classes_list.append(topk_classes)
|
| 33 |
+
|
| 34 |
+
topk_score = torch.stack(topk_score_list)
|
| 35 |
+
topk_inds = torch.stack(topk_inds_list)
|
| 36 |
+
topk_classes = torch.stack(topk_classes_list)
|
| 37 |
+
|
| 38 |
+
return topk_score, topk_inds, topk_classes
|
| 39 |
+
|
| 40 |
+
def gather_feat_idx(feats, inds, batch_size, batch_idx):
|
| 41 |
+
feats_list = []
|
| 42 |
+
dim = feats.size(-1)
|
| 43 |
+
_inds = inds.unsqueeze(-1).expand(inds.size(0), inds.size(1), dim)
|
| 44 |
+
|
| 45 |
+
for bs_idx in range(batch_size):
|
| 46 |
+
batch_inds = batch_idx==bs_idx
|
| 47 |
+
feat = feats[batch_inds]
|
| 48 |
+
feats_list.append(feat.gather(0, _inds[bs_idx]))
|
| 49 |
+
feats = torch.stack(feats_list)
|
| 50 |
+
return feats
|
| 51 |
+
|
| 52 |
+
|
| 53 |
+
def decode_bbox_from_voxels_nuscenes(batch_size, indices, obj, rot_cos, rot_sin,
|
| 54 |
+
center, center_z, dim, vel=None, iou=None, point_cloud_range=None, voxel_size=None, voxels_3d=None,
|
| 55 |
+
feature_map_stride=None, K=100, score_thresh=None, post_center_limit_range=None, add_features=None):
|
| 56 |
+
batch_idx = indices[:, 0]
|
| 57 |
+
spatial_indices = indices[:, 1:]
|
| 58 |
+
scores, inds, class_ids = _topk_1d(None, batch_size, batch_idx, obj, K=K, nuscenes=True)
|
| 59 |
+
|
| 60 |
+
center = gather_feat_idx(center, inds, batch_size, batch_idx)
|
| 61 |
+
rot_sin = gather_feat_idx(rot_sin, inds, batch_size, batch_idx)
|
| 62 |
+
rot_cos = gather_feat_idx(rot_cos, inds, batch_size, batch_idx)
|
| 63 |
+
center_z = gather_feat_idx(center_z, inds, batch_size, batch_idx)
|
| 64 |
+
dim = gather_feat_idx(dim, inds, batch_size, batch_idx)
|
| 65 |
+
spatial_indices = gather_feat_idx(spatial_indices, inds, batch_size, batch_idx)
|
| 66 |
+
|
| 67 |
+
if not add_features is None:
|
| 68 |
+
add_features = gather_feat_idx(add_features, inds, batch_size, batch_idx) #for add_feature in add_features]
|
| 69 |
+
|
| 70 |
+
if not isinstance(feature_map_stride, int):
|
| 71 |
+
feature_map_stride = gather_feat_idx(feature_map_stride.unsqueeze(-1), inds, batch_size, batch_idx)
|
| 72 |
+
|
| 73 |
+
angle = torch.atan2(rot_sin, rot_cos)
|
| 74 |
+
xs = (spatial_indices[:, :, -1:] + center[:, :, 0:1]) * feature_map_stride * voxel_size[0] + point_cloud_range[0]
|
| 75 |
+
ys = (spatial_indices[:, :, -2:-1] + center[:, :, 1:2]) * feature_map_stride * voxel_size[1] + point_cloud_range[1]
|
| 76 |
+
|
| 77 |
+
box_part_list = [xs, ys, center_z, dim, angle]
|
| 78 |
+
|
| 79 |
+
if not vel is None:
|
| 80 |
+
vel = gather_feat_idx(vel, inds, batch_size, batch_idx)
|
| 81 |
+
box_part_list.append(vel)
|
| 82 |
+
|
| 83 |
+
if not iou is None:
|
| 84 |
+
iou = gather_feat_idx(iou, inds, batch_size, batch_idx)
|
| 85 |
+
iou = torch.clamp(iou, min=0, max=1.)
|
| 86 |
+
|
| 87 |
+
final_box_preds = torch.cat((box_part_list), dim=-1)
|
| 88 |
+
final_scores = scores.view(batch_size, K)
|
| 89 |
+
final_class_ids = class_ids.view(batch_size, K)
|
| 90 |
+
if not add_features is None:
|
| 91 |
+
add_features = add_features.view(batch_size, K, add_features.shape[-1]) #for add_feature in add_features]
|
| 92 |
+
|
| 93 |
+
assert post_center_limit_range is not None
|
| 94 |
+
mask = (final_box_preds[..., :3] >= post_center_limit_range[:3]).all(2)
|
| 95 |
+
mask &= (final_box_preds[..., :3] <= post_center_limit_range[3:]).all(2)
|
| 96 |
+
|
| 97 |
+
if score_thresh is not None:
|
| 98 |
+
mask &= (final_scores > score_thresh)
|
| 99 |
+
|
| 100 |
+
ret_pred_dicts = []
|
| 101 |
+
for k in range(batch_size):
|
| 102 |
+
cur_mask = mask[k]
|
| 103 |
+
cur_boxes = final_box_preds[k, cur_mask]
|
| 104 |
+
cur_scores = final_scores[k, cur_mask]
|
| 105 |
+
cur_labels = final_class_ids[k, cur_mask]
|
| 106 |
+
cur_add_features = add_features[k, cur_mask] if not add_features is None else None
|
| 107 |
+
cur_iou = iou[k, cur_mask] if not iou is None else None
|
| 108 |
+
|
| 109 |
+
ret_pred_dicts.append({
|
| 110 |
+
'pred_boxes': cur_boxes,
|
| 111 |
+
'pred_scores': cur_scores,
|
| 112 |
+
'pred_labels': cur_labels,
|
| 113 |
+
'pred_ious': cur_iou,
|
| 114 |
+
'add_features': cur_add_features,
|
| 115 |
+
})
|
| 116 |
+
return ret_pred_dicts
|
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/config.py
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import yaml
|
| 2 |
+
from easydict import EasyDict
|
| 3 |
+
|
| 4 |
+
def merge_new_config(config, new_config):
|
| 5 |
+
if '_BASE_CONFIG_' in new_config:
|
| 6 |
+
with open(new_config['_BASE_CONFIG_'], 'r') as f:
|
| 7 |
+
try:
|
| 8 |
+
yaml_config = yaml.safe_load(f, Loader=yaml.FullLoader)
|
| 9 |
+
except:
|
| 10 |
+
yaml_config = yaml.safe_load(f)
|
| 11 |
+
config.update(EasyDict(yaml_config))
|
| 12 |
+
|
| 13 |
+
for key, val in new_config.items():
|
| 14 |
+
if not isinstance(val, dict):
|
| 15 |
+
config[key] = val
|
| 16 |
+
continue
|
| 17 |
+
if key not in config:
|
| 18 |
+
config[key] = EasyDict()
|
| 19 |
+
merge_new_config(config[key], val)
|
| 20 |
+
|
| 21 |
+
return config
|
| 22 |
+
|
| 23 |
+
def cfg_from_yaml_file(cfg_file, config):
|
| 24 |
+
with open(cfg_file, 'r') as f:
|
| 25 |
+
try:
|
| 26 |
+
new_config = yaml.safe_load(f, Loader=yaml.FullLoader)
|
| 27 |
+
except:
|
| 28 |
+
new_config = yaml.safe_load(f)
|
| 29 |
+
|
| 30 |
+
merge_new_config(config=config, new_config=new_config)
|
| 31 |
+
|
| 32 |
+
return config
|
| 33 |
+
|
| 34 |
+
cfg = EasyDict()
|
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/image_projection.py
ADDED
|
@@ -0,0 +1,103 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import numpy as np
|
| 3 |
+
import cv2
|
| 4 |
+
|
| 5 |
+
def get_data_info(info, cam_type):
|
| 6 |
+
|
| 7 |
+
cam_info = info[cam_type]
|
| 8 |
+
|
| 9 |
+
lidar2cam_r = np.linalg.inv(cam_info['sensor2lidar_rotation'])
|
| 10 |
+
lidar2cam_t = cam_info[
|
| 11 |
+
'sensor2lidar_translation'] @ lidar2cam_r.T
|
| 12 |
+
lidar2cam_rt = np.eye(4)
|
| 13 |
+
lidar2cam_rt[:3, :3] = lidar2cam_r.T
|
| 14 |
+
lidar2cam_rt[3, :3] = -lidar2cam_t
|
| 15 |
+
intrinsic = cam_info['cam_intrinsic']
|
| 16 |
+
viewpad = np.eye(4)
|
| 17 |
+
viewpad[:intrinsic.shape[0], :intrinsic.shape[1]] = intrinsic
|
| 18 |
+
lidar2img_rt = (viewpad @ lidar2cam_rt.T)
|
| 19 |
+
|
| 20 |
+
return lidar2img_rt
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
def _proj_voxel_image(voxel_coords, lidar2img_rt, voxel_size, point_cloud_range):
|
| 24 |
+
# voxel_coords [n ,4]
|
| 25 |
+
# lidar2img_rt [4, 4]
|
| 26 |
+
#x_input.indices [n, 4] [[0, Z, Y, Z]xn]
|
| 27 |
+
|
| 28 |
+
voxel_coords = voxel_coords[:, [3,2,1]]
|
| 29 |
+
device = voxel_coords.device
|
| 30 |
+
lidar2img_rt = torch.Tensor(lidar2img_rt).to(device)
|
| 31 |
+
# point_cloud_rangetensor([-51.2000, -51.2000, -5.0000, 51.2000, 51.2000, 3.0000])
|
| 32 |
+
voxel_coords = voxel_coords * voxel_size.unsqueeze(0) + point_cloud_range[:3].unsqueeze(0)
|
| 33 |
+
# (n, 4)
|
| 34 |
+
voxel_coords = torch.cat([voxel_coords, torch.ones((voxel_coords.shape[0], 1), device=device)], dim=-1)
|
| 35 |
+
points_image = torch.matmul(lidar2img_rt, voxel_coords.permute(1, 0)) #(voxel_coords @ lidar2img_rt).T
|
| 36 |
+
# (4, n)
|
| 37 |
+
depth = points_image[2:3] # (1, n)
|
| 38 |
+
points_image = points_image[:2] / torch.maximum(depth, torch.ones_like(depth*1e-4))
|
| 39 |
+
return points_image, depth
|
| 40 |
+
|
| 41 |
+
def _draw_image(points_image, image_path, depth):
|
| 42 |
+
image = cv2.imread(image_path)
|
| 43 |
+
points_image = points_image.int().cpu().numpy()
|
| 44 |
+
for i in range(points_image.shape[1]):
|
| 45 |
+
_point = points_image[:, i]
|
| 46 |
+
if _point[0] > 0 and _point[1] > 0 and depth[0][i] >0:
|
| 47 |
+
cv2.circle(image, tuple(_point), 1, (0,255,0), -1)
|
| 48 |
+
#cv2.imwrite("image.png", image)
|
| 49 |
+
return image
|
| 50 |
+
|
| 51 |
+
def _draw_mask(image_path, mask, color=None):
|
| 52 |
+
image = cv2.imread(image_path)
|
| 53 |
+
h, w, _ = image.shape
|
| 54 |
+
|
| 55 |
+
if color is None:
|
| 56 |
+
color = np.random.random(3)
|
| 57 |
+
|
| 58 |
+
image[mask] = image[mask] * color
|
| 59 |
+
#cv2.imwrite("image_mask.png", image)
|
| 60 |
+
return image
|
| 61 |
+
|
| 62 |
+
def _draw_3dbox(box, lidar2img_rt, image, mask=None, color=None, output_path="image_box.png"):
|
| 63 |
+
#image = cv2.imread(image_path)
|
| 64 |
+
h, w, _ = image.shape
|
| 65 |
+
|
| 66 |
+
if color is None:
|
| 67 |
+
color = np.random.random(3)
|
| 68 |
+
if not mask is None:
|
| 69 |
+
image[mask] = image[mask] * color
|
| 70 |
+
|
| 71 |
+
center_x, center_y, center_z, H, W, Z, angle = box[:7]
|
| 72 |
+
sin_angle, cos_angle = torch.sin(angle), torch.cos(angle)
|
| 73 |
+
top1 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
|
| 74 |
+
top2 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
|
| 75 |
+
top3 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
|
| 76 |
+
top4 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
|
| 77 |
+
|
| 78 |
+
down1 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
|
| 79 |
+
down2 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
|
| 80 |
+
down3 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
|
| 81 |
+
down4 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
|
| 82 |
+
points = torch.Tensor([top1, top2, top3, top4, down1, down2, down3, down4, [center_x, center_y, center_z]]) # (8, 3)
|
| 83 |
+
points = torch.cat([points, torch.ones((points.shape[0], 1), device=points.device)], dim=-1)
|
| 84 |
+
points_image = torch.matmul(torch.Tensor(lidar2img_rt).to(points.device), points.permute(1, 0))
|
| 85 |
+
depth = points_image[2:3] # (1, n)
|
| 86 |
+
points_image = points_image[:2] / torch.maximum(depth, torch.ones_like(depth*1e-4))
|
| 87 |
+
points_image = points_image.permute(1, 0).int().cpu().numpy() #(voxel_coords @ lidar2img_rt).T
|
| 88 |
+
lines = [[0, 1], [1, 2], [2, 3], [3, 0], [4, 5], [5, 6], [6, 7], [7, 4], [0, 4], [1, 5], [2, 6], [3, 7]]
|
| 89 |
+
|
| 90 |
+
cv2.circle(image, tuple(points_image[-1]), 3, (0, 255, 0), -1)
|
| 91 |
+
|
| 92 |
+
for line in lines:
|
| 93 |
+
cv2.line(image, tuple(points_image[line[0]]), tuple(points_image[line[1]]), tuple(color * 255), 2)
|
| 94 |
+
#cv2.imwrite(output_path, image)
|
| 95 |
+
return image
|
| 96 |
+
|
| 97 |
+
|
| 98 |
+
|
| 99 |
+
|
| 100 |
+
|
| 101 |
+
|
| 102 |
+
|
| 103 |
+
|
external/WildCamera/WildCamera/benchmark/benchmark_calibration.py
ADDED
|
@@ -0,0 +1,98 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from __future__ import print_function, division
|
| 2 |
+
import os, sys, inspect
|
| 3 |
+
project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
|
| 4 |
+
sys.path.insert(0, project_root)
|
| 5 |
+
|
| 6 |
+
import warnings
|
| 7 |
+
warnings.filterwarnings('ignore')
|
| 8 |
+
import torch
|
| 9 |
+
import argparse
|
| 10 |
+
from loguru import logger
|
| 11 |
+
from pprint import pprint
|
| 12 |
+
|
| 13 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
|
| 17 |
+
parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
|
| 18 |
+
parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
|
| 19 |
+
parser.add_argument('--experiment_name', type=str, help='name of the experiment', required=True, choices=['in_the_wild', 'gsv'])
|
| 20 |
+
|
| 21 |
+
def main_worker(args, wtassumption=False):
|
| 22 |
+
args.gpu = 0
|
| 23 |
+
pprint(vars(args))
|
| 24 |
+
|
| 25 |
+
model = NEWCRFIF(version='large07', pretrained=None)
|
| 26 |
+
model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
|
| 27 |
+
model.eval()
|
| 28 |
+
model.cuda()
|
| 29 |
+
logger.info("Load Model from %s" % args.load_ckpt)
|
| 30 |
+
|
| 31 |
+
if args.experiment_name == 'in_the_wild':
|
| 32 |
+
from WildCamera.evaluation.evaluate_intrinsic import EvaluateIntrinsic
|
| 33 |
+
evaluate_intrinsic = EvaluateIntrinsic()
|
| 34 |
+
evaluate_intrinsic.evaluate(model, args, steps=0, writer=None, group=None, wtassumption=wtassumption)
|
| 35 |
+
elif args.experiment_name == 'gsv':
|
| 36 |
+
from WildCamera.evaluation.evaluate_fov import EvaluateFov
|
| 37 |
+
evaluate_fov = EvaluateFov()
|
| 38 |
+
evaluate_fov.evaluate(model, args, steps=0, writer=None, group=None, wtassumption=wtassumption)
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
if __name__ == '__main__':
|
| 42 |
+
args = parser.parse_args()
|
| 43 |
+
|
| 44 |
+
args.world_size = torch.cuda.device_count()
|
| 45 |
+
args.augscale = 2.0
|
| 46 |
+
args.input_height = 480
|
| 47 |
+
args.input_width = 640
|
| 48 |
+
|
| 49 |
+
args.eval_workers = 4
|
| 50 |
+
args.l1_th = 0.02
|
| 51 |
+
|
| 52 |
+
if args.experiment_name == 'in_the_wild':
|
| 53 |
+
args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_all.pth')
|
| 54 |
+
args.datasets_train = [
|
| 55 |
+
'Nuscenes',
|
| 56 |
+
'KITTI',
|
| 57 |
+
'Cityscapes',
|
| 58 |
+
'NYUv2',
|
| 59 |
+
'ARKitScenes',
|
| 60 |
+
'MegaDepth',
|
| 61 |
+
'SUN3D',
|
| 62 |
+
'MVImgNet',
|
| 63 |
+
'Objectron',
|
| 64 |
+
]
|
| 65 |
+
args.datasets_eval = [
|
| 66 |
+
'Nuscenes',
|
| 67 |
+
'KITTI',
|
| 68 |
+
'Cityscapes',
|
| 69 |
+
'NYUv2',
|
| 70 |
+
'ARKitScenes',
|
| 71 |
+
'MegaDepth',
|
| 72 |
+
'SUN3D',
|
| 73 |
+
'MVImgNet',
|
| 74 |
+
'Objectron',
|
| 75 |
+
'Waymo',
|
| 76 |
+
'BIWIRGBDID',
|
| 77 |
+
'RGBD',
|
| 78 |
+
'ScanNet',
|
| 79 |
+
'CAD120',
|
| 80 |
+
'MVS',
|
| 81 |
+
'Scenes11',
|
| 82 |
+
]
|
| 83 |
+
main_worker(args, wtassumption=False)
|
| 84 |
+
main_worker(args, wtassumption=True)
|
| 85 |
+
elif args.experiment_name == 'gsv':
|
| 86 |
+
args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_gsv.pth')
|
| 87 |
+
args.datasets_train = [
|
| 88 |
+
'GSV'
|
| 89 |
+
]
|
| 90 |
+
args.datasets_eval = [
|
| 91 |
+
'GSV'
|
| 92 |
+
]
|
| 93 |
+
main_worker(args, wtassumption=False)
|
| 94 |
+
main_worker(args, wtassumption=True)
|
| 95 |
+
else:
|
| 96 |
+
raise NotImplementedError()
|
| 97 |
+
|
| 98 |
+
|
external/WildCamera/WildCamera/benchmark/benchmark_crop.py
ADDED
|
@@ -0,0 +1,63 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from __future__ import print_function, division
|
| 2 |
+
import os, sys, inspect, copy
|
| 3 |
+
project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
|
| 4 |
+
sys.path.insert(0, project_root)
|
| 5 |
+
|
| 6 |
+
import warnings
|
| 7 |
+
warnings.filterwarnings('ignore')
|
| 8 |
+
import torch
|
| 9 |
+
import argparse
|
| 10 |
+
from loguru import logger
|
| 11 |
+
from pprint import pprint
|
| 12 |
+
|
| 13 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 14 |
+
from WildCamera.evaluation.evaluate_crop import EvaluateCrop
|
| 15 |
+
|
| 16 |
+
parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
|
| 17 |
+
parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
|
| 18 |
+
parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
|
| 19 |
+
|
| 20 |
+
def main_worker(args):
|
| 21 |
+
args.gpu = 0
|
| 22 |
+
pprint(vars(args))
|
| 23 |
+
|
| 24 |
+
model = NEWCRFIF(version='large07', pretrained=None)
|
| 25 |
+
model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
|
| 26 |
+
model.eval()
|
| 27 |
+
model.cuda()
|
| 28 |
+
logger.info("Load Model from %s" % args.load_ckpt)
|
| 29 |
+
|
| 30 |
+
evaluate_crop = EvaluateCrop()
|
| 31 |
+
evaluate_crop.evaluate(model, args, group=None)
|
| 32 |
+
|
| 33 |
+
if __name__ == '__main__':
|
| 34 |
+
args = parser.parse_args()
|
| 35 |
+
|
| 36 |
+
args.world_size = 1
|
| 37 |
+
args.augscale = 2.0
|
| 38 |
+
args.input_height = 480
|
| 39 |
+
args.input_width = 640
|
| 40 |
+
args.eval_workers = 4
|
| 41 |
+
args.l1_th = 0.02
|
| 42 |
+
|
| 43 |
+
args.datasets_train = [
|
| 44 |
+
'KITTI',
|
| 45 |
+
'NYUv2',
|
| 46 |
+
'ARKitScenes',
|
| 47 |
+
'Waymo',
|
| 48 |
+
'RGBD',
|
| 49 |
+
'ScanNet',
|
| 50 |
+
'MVS',
|
| 51 |
+
]
|
| 52 |
+
args.datasets_eval = [
|
| 53 |
+
'KITTI',
|
| 54 |
+
'NYUv2',
|
| 55 |
+
'ARKitScenes',
|
| 56 |
+
'Waymo',
|
| 57 |
+
'RGBD',
|
| 58 |
+
'ScanNet',
|
| 59 |
+
'MVS',
|
| 60 |
+
]
|
| 61 |
+
|
| 62 |
+
args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_all.pth')
|
| 63 |
+
main_worker(args)
|
external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_megadepth.py
ADDED
|
@@ -0,0 +1,168 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, sys, inspect, pickle
|
| 2 |
+
project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
|
| 3 |
+
sys.path.insert(0, project_root)
|
| 4 |
+
|
| 5 |
+
import numpy as np
|
| 6 |
+
import torch
|
| 7 |
+
from tabulate import tabulate
|
| 8 |
+
from PIL import Image
|
| 9 |
+
from tqdm import tqdm
|
| 10 |
+
from argparse import ArgumentParser
|
| 11 |
+
from WildCamera.evaluation.evaluate_pose import compute_pose_error, pose_auc, estimate_pose, compute_relative_pose
|
| 12 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 13 |
+
|
| 14 |
+
class Megadepth1500Benchmark:
|
| 15 |
+
def __init__(self, data_root="data/megadepth", scene_names=None) -> None:
|
| 16 |
+
if scene_names is None:
|
| 17 |
+
self.scene_names = [
|
| 18 |
+
"0015_0.1_0.3.npz",
|
| 19 |
+
"0015_0.3_0.5.npz",
|
| 20 |
+
"0022_0.1_0.3.npz",
|
| 21 |
+
"0022_0.3_0.5.npz",
|
| 22 |
+
"0022_0.5_0.7.npz",
|
| 23 |
+
]
|
| 24 |
+
else:
|
| 25 |
+
self.scene_names = scene_names
|
| 26 |
+
self.scenes = [
|
| 27 |
+
np.load("{}/scene_info/{}".format(data_root, scene), allow_pickle=True)
|
| 28 |
+
for scene in self.scene_names
|
| 29 |
+
]
|
| 30 |
+
self.data_root = data_root
|
| 31 |
+
|
| 32 |
+
@torch.no_grad()
|
| 33 |
+
def benchmark(self, camera_calibrator, use_calibrated_intrinsic=False):
|
| 34 |
+
with torch.no_grad():
|
| 35 |
+
data_root = self.data_root
|
| 36 |
+
tot_e_t, tot_e_R, tot_e_pose = [], [], []
|
| 37 |
+
for scene_ind in range(len(self.scenes)):
|
| 38 |
+
scene = self.scenes[scene_ind]
|
| 39 |
+
pairs = scene["pair_infos"]
|
| 40 |
+
intrinsics = scene["intrinsics"]
|
| 41 |
+
poses = scene["poses"]
|
| 42 |
+
im_paths = scene["image_paths"]
|
| 43 |
+
pair_inds = range(len(pairs))
|
| 44 |
+
for pairind in tqdm(pair_inds, disable=False):
|
| 45 |
+
idx1, idx2 = pairs[pairind][0]
|
| 46 |
+
K1 = intrinsics[idx1].copy()
|
| 47 |
+
T1 = poses[idx1].copy()
|
| 48 |
+
R1, t1 = T1[:3, :3], T1[:3, 3]
|
| 49 |
+
K2 = intrinsics[idx2].copy()
|
| 50 |
+
T2 = poses[idx2].copy()
|
| 51 |
+
R2, t2 = T2[:3, :3], T2[:3, 3]
|
| 52 |
+
R, t = compute_relative_pose(R1, t1, R2, t2)
|
| 53 |
+
im1_path = f"{data_root}/{im_paths[idx1]}"
|
| 54 |
+
im2_path = f"{data_root}/{im_paths[idx2]}"
|
| 55 |
+
im1 = Image.open(im1_path)
|
| 56 |
+
w1, h1 = im1.size
|
| 57 |
+
im2 = Image.open(im2_path)
|
| 58 |
+
w2, h2 = im2.size
|
| 59 |
+
|
| 60 |
+
if not use_calibrated_intrinsic:
|
| 61 |
+
K1 = K1.copy()
|
| 62 |
+
K2 = K2.copy()
|
| 63 |
+
else:
|
| 64 |
+
K1, _ = camera_calibrator.inference(im1, wtassumption=True)
|
| 65 |
+
K1 = K1.astype(np.float64)
|
| 66 |
+
K2, _ = camera_calibrator.inference(im2, wtassumption=True)
|
| 67 |
+
K2 = K2.astype(np.float64)
|
| 68 |
+
|
| 69 |
+
scale1 = 1200 / max(w1, h1)
|
| 70 |
+
scale2 = 1200 / max(w2, h2)
|
| 71 |
+
w1, h1 = scale1 * w1, scale1 * h1
|
| 72 |
+
w2, h2 = scale2 * w2, scale2 * h2
|
| 73 |
+
K1[:2] = K1[:2] * scale1
|
| 74 |
+
K2[:2] = K2[:2] * scale2
|
| 75 |
+
|
| 76 |
+
corres_fold = os.path.join(self.data_root, 'MegaDepthCorrespondence')
|
| 77 |
+
sv_path = os.path.join(corres_fold, '{}_{}.pkl'.format(str(scene_ind), str(pairind)))
|
| 78 |
+
with open(sv_path, 'rb') as f:
|
| 79 |
+
sparse_matches = pickle.load(f)
|
| 80 |
+
|
| 81 |
+
kpts1 = sparse_matches[:, :2]
|
| 82 |
+
kpts1 = (
|
| 83 |
+
np.stack(
|
| 84 |
+
(
|
| 85 |
+
w1 * (kpts1[:, 0] + 1) / 2,
|
| 86 |
+
h1 * (kpts1[:, 1] + 1) / 2,
|
| 87 |
+
),
|
| 88 |
+
axis=-1,
|
| 89 |
+
)
|
| 90 |
+
)
|
| 91 |
+
kpts2 = sparse_matches[:, 2:]
|
| 92 |
+
kpts2 = (
|
| 93 |
+
np.stack(
|
| 94 |
+
(
|
| 95 |
+
w2 * (kpts2[:, 0] + 1) / 2,
|
| 96 |
+
h2 * (kpts2[:, 1] + 1) / 2,
|
| 97 |
+
),
|
| 98 |
+
axis=-1,
|
| 99 |
+
)
|
| 100 |
+
)
|
| 101 |
+
|
| 102 |
+
for _ in range(5):
|
| 103 |
+
shuffling = np.random.permutation(np.arange(len(kpts1)))
|
| 104 |
+
kpts1 = kpts1[shuffling]
|
| 105 |
+
kpts2 = kpts2[shuffling]
|
| 106 |
+
try:
|
| 107 |
+
norm_threshold = 0.5 / (
|
| 108 |
+
np.mean(np.abs(K1[:2, :2])) + np.mean(np.abs(K2[:2, :2])))
|
| 109 |
+
R_est, t_est, mask = estimate_pose(
|
| 110 |
+
kpts1,
|
| 111 |
+
kpts2,
|
| 112 |
+
K1,
|
| 113 |
+
K2,
|
| 114 |
+
norm_threshold,
|
| 115 |
+
conf=0.99999
|
| 116 |
+
)
|
| 117 |
+
T1_to_2_est = np.concatenate((R_est, t_est), axis=-1) #
|
| 118 |
+
e_t, e_R = compute_pose_error(T1_to_2_est, R, t)
|
| 119 |
+
e_pose = max(e_t, e_R)
|
| 120 |
+
except Exception as e:
|
| 121 |
+
e_t, e_R = 90, 90
|
| 122 |
+
e_pose = max(e_t, e_R)
|
| 123 |
+
|
| 124 |
+
tot_e_t.append(e_t)
|
| 125 |
+
tot_e_R.append(e_R)
|
| 126 |
+
tot_e_pose.append(e_pose)
|
| 127 |
+
|
| 128 |
+
tot_e_pose = np.array(tot_e_pose)
|
| 129 |
+
thresholds = [5, 10, 20]
|
| 130 |
+
auc = pose_auc(tot_e_pose, thresholds)
|
| 131 |
+
acc_5 = (tot_e_pose < 5).mean()
|
| 132 |
+
acc_10 = (tot_e_pose < 10).mean()
|
| 133 |
+
acc_15 = (tot_e_pose < 15).mean()
|
| 134 |
+
acc_20 = (tot_e_pose < 20).mean()
|
| 135 |
+
map_5 = acc_5
|
| 136 |
+
map_10 = np.mean([acc_5, acc_10])
|
| 137 |
+
map_20 = np.mean([acc_5, acc_10, acc_15, acc_20])
|
| 138 |
+
|
| 139 |
+
result = {
|
| 140 |
+
"auc_5": auc[0],
|
| 141 |
+
"auc_10": auc[1],
|
| 142 |
+
"auc_20": auc[2],
|
| 143 |
+
"map_5": map_5,
|
| 144 |
+
"map_10": map_10,
|
| 145 |
+
"map_20": map_20,
|
| 146 |
+
}
|
| 147 |
+
|
| 148 |
+
for k in result.keys():
|
| 149 |
+
result[k] = result[k] * 100
|
| 150 |
+
|
| 151 |
+
return result
|
| 152 |
+
|
| 153 |
+
if __name__ == "__main__":
|
| 154 |
+
parser = ArgumentParser()
|
| 155 |
+
parser.add_argument("--data_path", type=str, required=True)
|
| 156 |
+
parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
|
| 157 |
+
args, _ = parser.parse_known_args()
|
| 158 |
+
|
| 159 |
+
megaloftr_benchmark = Megadepth1500Benchmark(args.data_path)
|
| 160 |
+
|
| 161 |
+
args.load_ckpt = os.path.join(project_root, 'model_zoo', 'Release', 'wild_camera_all.pth')
|
| 162 |
+
camera_calibrator = NEWCRFIF(version='large07', pretrained=None)
|
| 163 |
+
camera_calibrator.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
|
| 164 |
+
camera_calibrator.eval()
|
| 165 |
+
camera_calibrator.cuda()
|
| 166 |
+
|
| 167 |
+
result = megaloftr_benchmark.benchmark(camera_calibrator, use_calibrated_intrinsic=True)
|
| 168 |
+
print(tabulate(result.items(), headers=['Metric', 'Scores'], tablefmt='fancy_grid', floatfmt=".2f", numalign="left"))
|
external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_scannet.py
ADDED
|
@@ -0,0 +1,165 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, sys, inspect, tqdm, copy, pickle
|
| 2 |
+
project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
|
| 3 |
+
sys.path.insert(0, project_root)
|
| 4 |
+
|
| 5 |
+
import numpy as np
|
| 6 |
+
import torch
|
| 7 |
+
from tabulate import tabulate
|
| 8 |
+
from PIL import Image
|
| 9 |
+
from tqdm import tqdm
|
| 10 |
+
from argparse import ArgumentParser
|
| 11 |
+
from WildCamera.evaluation.evaluate_pose import compute_pose_error, pose_auc, estimate_pose, compute_relative_pose
|
| 12 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 13 |
+
|
| 14 |
+
class ScanNetBenchmark:
|
| 15 |
+
def __init__(self, data_root="data/scannet") -> None:
|
| 16 |
+
self.data_root = data_root
|
| 17 |
+
|
| 18 |
+
def benchmark(self, camera_calibrator, use_calibrated_intrinsic=False):
|
| 19 |
+
with torch.no_grad():
|
| 20 |
+
data_root = self.data_root
|
| 21 |
+
tmp = np.load(os.path.join(data_root, "test.npz"))
|
| 22 |
+
pairs, rel_pose = tmp["name"], tmp["rel_pose"]
|
| 23 |
+
tot_e_t, tot_e_R, tot_e_pose = [], [], []
|
| 24 |
+
np.random.seed(0)
|
| 25 |
+
pair_inds = np.random.choice(
|
| 26 |
+
range(len(pairs)), size=len(pairs), replace=False
|
| 27 |
+
)
|
| 28 |
+
for cnt, pairind in enumerate(tqdm(pair_inds)):
|
| 29 |
+
scene = pairs[pairind]
|
| 30 |
+
scene_name = f"scene0{scene[0]}_00"
|
| 31 |
+
im1_path = os.path.join(
|
| 32 |
+
self.data_root,
|
| 33 |
+
"scans_test",
|
| 34 |
+
scene_name,
|
| 35 |
+
"color",
|
| 36 |
+
f"{scene[2]}.jpg",
|
| 37 |
+
)
|
| 38 |
+
im1 = Image.open(im1_path)
|
| 39 |
+
im2_path = os.path.join(
|
| 40 |
+
self.data_root,
|
| 41 |
+
"scans_test",
|
| 42 |
+
scene_name,
|
| 43 |
+
"color",
|
| 44 |
+
f"{scene[3]}.jpg",
|
| 45 |
+
)
|
| 46 |
+
im2 = Image.open(im2_path)
|
| 47 |
+
T_gt = rel_pose[pairind].reshape(3, 4)
|
| 48 |
+
R, t = T_gt[:3, :3], T_gt[:3, 3]
|
| 49 |
+
K = np.stack(
|
| 50 |
+
[
|
| 51 |
+
np.array([float(i) for i in r.split()])
|
| 52 |
+
for r in open(
|
| 53 |
+
os.path.join(
|
| 54 |
+
self.data_root,
|
| 55 |
+
"scans_test",
|
| 56 |
+
scene_name,
|
| 57 |
+
"intrinsic",
|
| 58 |
+
"intrinsic_color.txt",
|
| 59 |
+
),
|
| 60 |
+
"r",
|
| 61 |
+
)
|
| 62 |
+
.read()
|
| 63 |
+
.split("\n")
|
| 64 |
+
if r
|
| 65 |
+
]
|
| 66 |
+
)
|
| 67 |
+
w1, h1 = im1.size
|
| 68 |
+
w2, h2 = im2.size
|
| 69 |
+
|
| 70 |
+
if not use_calibrated_intrinsic:
|
| 71 |
+
K1 = K.copy()
|
| 72 |
+
K2 = K.copy()
|
| 73 |
+
else:
|
| 74 |
+
K1_, _ = camera_calibrator.inference(im1, wtassumption=True)
|
| 75 |
+
K1_ = K1_.astype(np.float64)
|
| 76 |
+
K2_, _ = camera_calibrator.inference(im2, wtassumption=True)
|
| 77 |
+
K2_ = K2_.astype(np.float64)
|
| 78 |
+
|
| 79 |
+
K1 = np.eye(4)
|
| 80 |
+
K1[0:3, 0:3] = K1_
|
| 81 |
+
|
| 82 |
+
K2 = np.eye(4)
|
| 83 |
+
K2[0:3, 0:3] = K2_
|
| 84 |
+
|
| 85 |
+
scale1 = 480 / min(w1, h1)
|
| 86 |
+
scale2 = 480 / min(w2, h2)
|
| 87 |
+
w1, h1 = scale1 * w1, scale1 * h1
|
| 88 |
+
w2, h2 = scale2 * w2, scale2 * h2
|
| 89 |
+
K1 = K1 * scale1
|
| 90 |
+
K2 = K2 * scale2
|
| 91 |
+
|
| 92 |
+
corres_fold = os.path.join(self.data_root, 'ScanNetCorrespondence')
|
| 93 |
+
sv_path = os.path.join(corres_fold, '{}.pkl'.format(str(pairind)))
|
| 94 |
+
with open(sv_path, 'rb') as f:
|
| 95 |
+
sparse_matches = pickle.load(f)
|
| 96 |
+
|
| 97 |
+
kpts1 = sparse_matches[:, :2]
|
| 98 |
+
kpts2 = sparse_matches[:, 2:]
|
| 99 |
+
|
| 100 |
+
for _ in range(5):
|
| 101 |
+
shuffling = np.random.permutation(np.arange(len(kpts1)))
|
| 102 |
+
kpts1 = kpts1[shuffling]
|
| 103 |
+
kpts2 = kpts2[shuffling]
|
| 104 |
+
try:
|
| 105 |
+
norm_threshold = 0.8 / (
|
| 106 |
+
np.mean(np.abs(K1[:2, :2])) + np.mean(np.abs(K2[:2, :2])))
|
| 107 |
+
R_est, t_est, mask = estimate_pose(
|
| 108 |
+
kpts1,
|
| 109 |
+
kpts2,
|
| 110 |
+
K1,
|
| 111 |
+
K2,
|
| 112 |
+
norm_threshold,
|
| 113 |
+
conf=0.99999,
|
| 114 |
+
)
|
| 115 |
+
T1_to_2_est = np.concatenate((R_est, t_est), axis=-1) #
|
| 116 |
+
e_t, e_R = compute_pose_error(T1_to_2_est, R, t)
|
| 117 |
+
e_pose = max(e_t, e_R)
|
| 118 |
+
except Exception as _:
|
| 119 |
+
e_t, e_R = 90, 90
|
| 120 |
+
e_pose = max(e_t, e_R)
|
| 121 |
+
tot_e_t.append(e_t)
|
| 122 |
+
tot_e_R.append(e_R)
|
| 123 |
+
tot_e_pose.append(e_pose)
|
| 124 |
+
|
| 125 |
+
tot_e_pose = np.array(tot_e_pose)
|
| 126 |
+
thresholds = [5, 10, 20]
|
| 127 |
+
auc = pose_auc(tot_e_pose, thresholds)
|
| 128 |
+
acc_5 = (tot_e_pose < 5).mean()
|
| 129 |
+
acc_10 = (tot_e_pose < 10).mean()
|
| 130 |
+
acc_15 = (tot_e_pose < 15).mean()
|
| 131 |
+
acc_20 = (tot_e_pose < 20).mean()
|
| 132 |
+
map_5 = acc_5
|
| 133 |
+
map_10 = np.mean([acc_5, acc_10])
|
| 134 |
+
map_20 = np.mean([acc_5, acc_10, acc_15, acc_20])
|
| 135 |
+
|
| 136 |
+
result = {
|
| 137 |
+
"auc_5": auc[0],
|
| 138 |
+
"auc_10": auc[1],
|
| 139 |
+
"auc_20": auc[2],
|
| 140 |
+
"map_5": map_5,
|
| 141 |
+
"map_10": map_10,
|
| 142 |
+
"map_20": map_20,
|
| 143 |
+
}
|
| 144 |
+
|
| 145 |
+
for k in result.keys():
|
| 146 |
+
result[k] = result[k] * 100
|
| 147 |
+
|
| 148 |
+
return result
|
| 149 |
+
|
| 150 |
+
if __name__ == "__main__":
|
| 151 |
+
parser = ArgumentParser()
|
| 152 |
+
parser.add_argument("--data_path", type=str, required=True)
|
| 153 |
+
parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
|
| 154 |
+
args, _ = parser.parse_known_args()
|
| 155 |
+
|
| 156 |
+
scannet_benchmark = ScanNetBenchmark(args.data_path)
|
| 157 |
+
|
| 158 |
+
args.load_ckpt = os.path.join(project_root, 'model_zoo', 'Release', 'wild_camera_all.pth')
|
| 159 |
+
camera_calibrator = NEWCRFIF(version='large07', pretrained=None)
|
| 160 |
+
camera_calibrator.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
|
| 161 |
+
camera_calibrator.eval()
|
| 162 |
+
camera_calibrator.cuda()
|
| 163 |
+
|
| 164 |
+
result = scannet_benchmark.benchmark(camera_calibrator, use_calibrated_intrinsic=True)
|
| 165 |
+
print(tabulate(result.items(), headers=['Metric', 'Scores'], tablefmt='fancy_grid', floatfmt=".2f", numalign="left"))
|
external/WildCamera/WildCamera/datasets/GSV.py
ADDED
|
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, io, glob, natsort, random, copy
|
| 2 |
+
import h5py
|
| 3 |
+
import torch
|
| 4 |
+
import numpy as np
|
| 5 |
+
from PIL import Image
|
| 6 |
+
from torchvision import transforms
|
| 7 |
+
|
| 8 |
+
class GSV:
|
| 9 |
+
def __init__(self, data_root, ht=384, wt=512, shuffleseed=None, split='train') -> None:
|
| 10 |
+
split_path = os.path.join('splits', 'gsv_{}.txt'.format(split))
|
| 11 |
+
|
| 12 |
+
with open(split_path) as file:
|
| 13 |
+
data_names = [line.rstrip() for line in file]
|
| 14 |
+
|
| 15 |
+
if split == 'train':
|
| 16 |
+
if shuffleseed is not None:
|
| 17 |
+
random.seed(shuffleseed)
|
| 18 |
+
data_names = copy.deepcopy(data_names) + copy.deepcopy(data_names)
|
| 19 |
+
random.shuffle(data_names)
|
| 20 |
+
|
| 21 |
+
self.data_root = data_root
|
| 22 |
+
self.wt, self.ht = wt, ht
|
| 23 |
+
self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
|
| 24 |
+
self.tensor = transforms.ToTensor()
|
| 25 |
+
self.data_names = data_names
|
| 26 |
+
|
| 27 |
+
def __len__(self):
|
| 28 |
+
return len(self.data_names)
|
| 29 |
+
|
| 30 |
+
def load_im(self, im_ref):
|
| 31 |
+
im = Image.open(im_ref)
|
| 32 |
+
return im
|
| 33 |
+
|
| 34 |
+
def __getitem__(self, idx):
|
| 35 |
+
# read intrinsics of original size
|
| 36 |
+
scene_name, entry_name = self.data_names[idx].split(' ')
|
| 37 |
+
h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
|
| 38 |
+
|
| 39 |
+
with h5py.File(h5pypath, 'r') as hf:
|
| 40 |
+
K_color = np.array(hf['intrinsic'][entry_name])
|
| 41 |
+
|
| 42 |
+
# Load positive pair data
|
| 43 |
+
rgb = self.load_im(io.BytesIO(np.array(hf['color'][entry_name])))
|
| 44 |
+
w, h = rgb.size
|
| 45 |
+
rgb = rgb.resize((self.wt, self.ht))
|
| 46 |
+
|
| 47 |
+
scaleM = np.eye(3)
|
| 48 |
+
scaleM[0, 0] = self.wt / w
|
| 49 |
+
scaleM[1, 1] = self.ht / h
|
| 50 |
+
aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
|
| 51 |
+
|
| 52 |
+
K = torch.from_numpy(scaleM @ K_color).float()
|
| 53 |
+
|
| 54 |
+
fovy = 2 * np.arctan(h / K_color[1, 1] / 2)
|
| 55 |
+
fovy = fovy.item()
|
| 56 |
+
|
| 57 |
+
# Recompute camera intrinsic matrix due to the resize
|
| 58 |
+
rgb = self.normalize(self.tensor(rgb))
|
| 59 |
+
|
| 60 |
+
data_dict = {
|
| 61 |
+
'K': K,
|
| 62 |
+
'rgb': rgb,
|
| 63 |
+
'K_raw': torch.clone(K),
|
| 64 |
+
'rgb_raw': torch.clone(rgb),
|
| 65 |
+
'aspect_ratio_restoration': aspect_ratio_restoration,
|
| 66 |
+
'fovy': fovy,
|
| 67 |
+
# Only Required in Crop Evaluation and Visualization in Training
|
| 68 |
+
'T': torch.eye(3, dtype=torch.float32),
|
| 69 |
+
'scaleM': scaleM.astype(np.float32),
|
| 70 |
+
'size_wo_change': (h, w),
|
| 71 |
+
}
|
| 72 |
+
|
| 73 |
+
return data_dict
|
external/WildCamera/WildCamera/datasets/GenericDataset.py
ADDED
|
@@ -0,0 +1,171 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, io, glob, natsort, random, copy
|
| 2 |
+
import h5py
|
| 3 |
+
import torch
|
| 4 |
+
import numpy as np
|
| 5 |
+
import hashlib
|
| 6 |
+
from PIL import Image
|
| 7 |
+
from torchvision import transforms
|
| 8 |
+
from tools.tools import coords_gridN, resample_rgb, apply_augmentation
|
| 9 |
+
from torchvision.transforms import ColorJitter
|
| 10 |
+
|
| 11 |
+
def add_white_noise(rgb):
|
| 12 |
+
w, h = rgb.size
|
| 13 |
+
rgb = np.array(rgb).astype(np.float32)
|
| 14 |
+
rgb = rgb + np.random.randint(9, size=(h, w, 3)) - 2
|
| 15 |
+
rgb = np.clip(rgb, a_min=0, a_max=255)
|
| 16 |
+
rgb = np.round(rgb).astype(np.uint8)
|
| 17 |
+
rgb = Image.fromarray(rgb)
|
| 18 |
+
return rgb
|
| 19 |
+
|
| 20 |
+
class GenericDataset:
|
| 21 |
+
def __init__(self,
|
| 22 |
+
data_root,
|
| 23 |
+
ht=384, wt=512,
|
| 24 |
+
augmentation=False,
|
| 25 |
+
shuffleseed=None,
|
| 26 |
+
split='train',
|
| 27 |
+
datasetname='MegaDepth',
|
| 28 |
+
augscale=2.0,
|
| 29 |
+
no_change_prob=0.1,
|
| 30 |
+
coloraugmentation=False,
|
| 31 |
+
coloraugmentation_scale=0.1,
|
| 32 |
+
transformcategory='transform_calibration'
|
| 33 |
+
) -> None:
|
| 34 |
+
|
| 35 |
+
name_mapping = {
|
| 36 |
+
'ScanNet': 'scannet',
|
| 37 |
+
'MegaDepth': 'megadepth',
|
| 38 |
+
'NYUv2': 'nyuv2',
|
| 39 |
+
'Cityscapes': 'cityscapes',
|
| 40 |
+
'MVS': 'mvs',
|
| 41 |
+
'RGBD': 'rgbd',
|
| 42 |
+
'Scenes11': 'scenes11',
|
| 43 |
+
'SUN3D': 'sun3d',
|
| 44 |
+
'BIWIRGBDID': 'biwirgbdid',
|
| 45 |
+
'CAD120': 'cad120',
|
| 46 |
+
'KITTI': 'kitti',
|
| 47 |
+
'Waymo': 'waymo',
|
| 48 |
+
'Nuscenes': 'nuscenes',
|
| 49 |
+
'ARKitScenes': 'arkitscenes',
|
| 50 |
+
'Objectron': 'objectron',
|
| 51 |
+
'MVImgNet': 'mvimgnet'
|
| 52 |
+
}
|
| 53 |
+
split_path = os.path.join('splits', '{}_{}.txt'.format(name_mapping[datasetname], split))
|
| 54 |
+
with open(split_path) as file:
|
| 55 |
+
data_names = [line.rstrip() for line in file]
|
| 56 |
+
|
| 57 |
+
if split == 'train':
|
| 58 |
+
if shuffleseed is not None:
|
| 59 |
+
random.seed(shuffleseed)
|
| 60 |
+
random.shuffle(data_names)
|
| 61 |
+
self.training = True
|
| 62 |
+
else:
|
| 63 |
+
self.training = False
|
| 64 |
+
|
| 65 |
+
self.data_root = data_root
|
| 66 |
+
|
| 67 |
+
self.wt, self.ht = wt, ht
|
| 68 |
+
self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
|
| 69 |
+
self.tensor = transforms.ToTensor()
|
| 70 |
+
|
| 71 |
+
self.data_names = data_names
|
| 72 |
+
self.augmentation = augmentation
|
| 73 |
+
self.augscale = augscale
|
| 74 |
+
self.no_change_prob = no_change_prob
|
| 75 |
+
self.coloraugmentation = coloraugmentation
|
| 76 |
+
self.coloraugmentation_scale = coloraugmentation_scale
|
| 77 |
+
|
| 78 |
+
self.datasetname = datasetname
|
| 79 |
+
self.transformcategory = transformcategory
|
| 80 |
+
|
| 81 |
+
def __len__(self):
|
| 82 |
+
return len(self.data_names)
|
| 83 |
+
|
| 84 |
+
def load_im(self, im_ref):
|
| 85 |
+
im = Image.open(im_ref)
|
| 86 |
+
return im
|
| 87 |
+
|
| 88 |
+
def color_augmentation_fun(self, rgb):
|
| 89 |
+
if random.uniform(0, 1) > 0.5:
|
| 90 |
+
colorjitter = ColorJitter(
|
| 91 |
+
brightness=self.coloraugmentation_scale,
|
| 92 |
+
contrast=self.coloraugmentation_scale,
|
| 93 |
+
saturation=self.coloraugmentation_scale,
|
| 94 |
+
hue=self.coloraugmentation_scale / 3.14
|
| 95 |
+
)
|
| 96 |
+
rgb = colorjitter(rgb)
|
| 97 |
+
rgb = add_white_noise(rgb)
|
| 98 |
+
return rgb
|
| 99 |
+
|
| 100 |
+
def __getitem__(self, idx):
|
| 101 |
+
# While augmenting novel intrinsic, we follow:
|
| 102 |
+
# Step 1 : Resize from original resolution to 480 x 640 (input height x input width)
|
| 103 |
+
# Step 2 : Apply random spatial augmentation
|
| 104 |
+
|
| 105 |
+
scene_name, stem_name = self.data_names[idx].split(' ')
|
| 106 |
+
h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
|
| 107 |
+
|
| 108 |
+
if not os.path.exists(h5pypath):
|
| 109 |
+
print("H5 file %s missing" % h5pypath)
|
| 110 |
+
assert os.path.exists(h5pypath)
|
| 111 |
+
|
| 112 |
+
with h5py.File(h5pypath, 'r') as hf:
|
| 113 |
+
# Load Intrinsic
|
| 114 |
+
K_color = np.array(hf['intrinsic'][stem_name])
|
| 115 |
+
|
| 116 |
+
# Load RGB
|
| 117 |
+
rgb = self.load_im(io.BytesIO(np.array(hf['color'][stem_name])))
|
| 118 |
+
|
| 119 |
+
w, h = rgb.size
|
| 120 |
+
|
| 121 |
+
# Step 1 : Resize
|
| 122 |
+
rgb = rgb.resize((self.wt, self.ht))
|
| 123 |
+
|
| 124 |
+
scaleM = np.eye(3)
|
| 125 |
+
scaleM[0, 0] = self.wt / w
|
| 126 |
+
scaleM[1, 1] = self.ht / h
|
| 127 |
+
aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
|
| 128 |
+
|
| 129 |
+
K = torch.from_numpy(scaleM @ K_color).float()
|
| 130 |
+
|
| 131 |
+
# Color Augmentation only in training
|
| 132 |
+
rgb = self.color_augmentation_fun(rgb) if self.coloraugmentation else rgb
|
| 133 |
+
|
| 134 |
+
# Normalization
|
| 135 |
+
rgb = self.normalize(self.tensor(rgb))
|
| 136 |
+
|
| 137 |
+
# Save RAW
|
| 138 |
+
K_raw = torch.clone(K)
|
| 139 |
+
rgb_raw = torch.clone(rgb)
|
| 140 |
+
|
| 141 |
+
# Step 2 : Random spatial augmentation
|
| 142 |
+
if self.augmentation:
|
| 143 |
+
if self.training:
|
| 144 |
+
rgb, K, T = apply_augmentation(
|
| 145 |
+
rgb, K, seed=None, augscale=self.augscale, no_change_prob=self.no_change_prob
|
| 146 |
+
)
|
| 147 |
+
else:
|
| 148 |
+
T = np.array(h5py.File(h5pypath, 'r')[self.transformcategory][stem_name])
|
| 149 |
+
T = torch.from_numpy(T).float()
|
| 150 |
+
K = torch.inverse(T) @ K
|
| 151 |
+
|
| 152 |
+
_, h_, w_ = rgb.shape
|
| 153 |
+
rgb = resample_rgb(rgb.unsqueeze(0), T, 1, h_, w_, rgb.device).squeeze(0)
|
| 154 |
+
else:
|
| 155 |
+
T = torch.eye(3, dtype=torch.float32)
|
| 156 |
+
|
| 157 |
+
# Exportation
|
| 158 |
+
data_dict = {
|
| 159 |
+
'K': K,
|
| 160 |
+
'rgb': rgb,
|
| 161 |
+
'K_raw': K_raw,
|
| 162 |
+
'rgb_raw': rgb_raw,
|
| 163 |
+
'aspect_ratio_restoration': aspect_ratio_restoration,
|
| 164 |
+
'datasetname': self.datasetname,
|
| 165 |
+
# Only Required in Crop Evaluation
|
| 166 |
+
'T': T,
|
| 167 |
+
'scaleM': scaleM.astype(np.float32),
|
| 168 |
+
'size_wo_change': (h, w),
|
| 169 |
+
}
|
| 170 |
+
|
| 171 |
+
return data_dict
|
external/WildCamera/WildCamera/datasets/IncdDataset.py
ADDED
|
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import torch
|
| 3 |
+
import numpy as np
|
| 4 |
+
|
| 5 |
+
from WildCamera.datasets.MegaDepth import MegaDepth
|
| 6 |
+
from WildCamera.datasets.GSV import GSV
|
| 7 |
+
from WildCamera.datasets.GenericDataset import GenericDataset
|
| 8 |
+
|
| 9 |
+
from torch.utils.data import Subset
|
| 10 |
+
|
| 11 |
+
def IncdDataset(
|
| 12 |
+
data_root,
|
| 13 |
+
datasets_included,
|
| 14 |
+
ht=384, wt=512,
|
| 15 |
+
augmentation=False,
|
| 16 |
+
shuffleseed=None,
|
| 17 |
+
split='train',
|
| 18 |
+
dataset_favour_long=0.0,
|
| 19 |
+
augscale=2.0,
|
| 20 |
+
no_change_prob=0.1,
|
| 21 |
+
coloraugmentation=False,
|
| 22 |
+
coloraugmentation_scale=0.0,
|
| 23 |
+
transformcategory='transform_crop' # For Evaluaiton Purpose
|
| 24 |
+
) -> None:
|
| 25 |
+
datasets = dict()
|
| 26 |
+
|
| 27 |
+
for datasetname in datasets_included:
|
| 28 |
+
if datasetname == 'GSV':
|
| 29 |
+
datasets[datasetname] = GSV(
|
| 30 |
+
os.path.join(data_root, 'GSV'),
|
| 31 |
+
ht=ht, wt=wt,
|
| 32 |
+
shuffleseed=shuffleseed,
|
| 33 |
+
split=split
|
| 34 |
+
)
|
| 35 |
+
elif datasetname == 'MegaDepth':
|
| 36 |
+
datasets[datasetname] = MegaDepth(
|
| 37 |
+
os.path.join(data_root, 'MegaDepth'),
|
| 38 |
+
ht=ht, wt=wt,
|
| 39 |
+
shuffleseed=shuffleseed,
|
| 40 |
+
split=split,
|
| 41 |
+
)
|
| 42 |
+
elif datasetname == 'BIWIRGBDID':
|
| 43 |
+
datasets[datasetname] = GenericDataset(
|
| 44 |
+
os.path.join(data_root, 'BIWIRGBDID'),
|
| 45 |
+
ht=ht, wt=wt,
|
| 46 |
+
augmentation=False,
|
| 47 |
+
shuffleseed=shuffleseed,
|
| 48 |
+
split=split,
|
| 49 |
+
datasetname=datasetname,
|
| 50 |
+
coloraugmentation=coloraugmentation,
|
| 51 |
+
coloraugmentation_scale=coloraugmentation_scale
|
| 52 |
+
)
|
| 53 |
+
elif datasetname == 'CAD120':
|
| 54 |
+
datasets[datasetname] = GenericDataset(
|
| 55 |
+
os.path.join(data_root, 'CAD120'),
|
| 56 |
+
ht=ht, wt=wt,
|
| 57 |
+
augmentation=False,
|
| 58 |
+
shuffleseed=shuffleseed,
|
| 59 |
+
split=split,
|
| 60 |
+
datasetname=datasetname,
|
| 61 |
+
coloraugmentation=coloraugmentation,
|
| 62 |
+
coloraugmentation_scale=coloraugmentation_scale
|
| 63 |
+
)
|
| 64 |
+
elif (datasetname == 'Objectron') or (datasetname == 'MVImgNet'):
|
| 65 |
+
augscale_obj = 1 + (augscale - 1) / 5
|
| 66 |
+
datasets[datasetname] = GenericDataset(
|
| 67 |
+
os.path.join(data_root, datasetname),
|
| 68 |
+
ht=ht, wt=wt,
|
| 69 |
+
augmentation=True,
|
| 70 |
+
shuffleseed=shuffleseed,
|
| 71 |
+
split=split,
|
| 72 |
+
datasetname=datasetname,
|
| 73 |
+
augscale=augscale_obj,
|
| 74 |
+
no_change_prob=no_change_prob,
|
| 75 |
+
coloraugmentation=coloraugmentation,
|
| 76 |
+
coloraugmentation_scale=coloraugmentation_scale
|
| 77 |
+
)
|
| 78 |
+
else:
|
| 79 |
+
datasets[datasetname] = GenericDataset(
|
| 80 |
+
os.path.join(data_root, datasetname),
|
| 81 |
+
ht=ht, wt=wt,
|
| 82 |
+
augmentation=augmentation,
|
| 83 |
+
shuffleseed=shuffleseed,
|
| 84 |
+
split=split,
|
| 85 |
+
datasetname=datasetname,
|
| 86 |
+
augscale=augscale,
|
| 87 |
+
no_change_prob=no_change_prob,
|
| 88 |
+
coloraugmentation=coloraugmentation,
|
| 89 |
+
coloraugmentation_scale=coloraugmentation_scale,
|
| 90 |
+
transformcategory=transformcategory
|
| 91 |
+
)
|
| 92 |
+
|
| 93 |
+
if split == 'train':
|
| 94 |
+
min_len = np.array([len(datasets[key]) for key in datasets.keys()]).min()
|
| 95 |
+
|
| 96 |
+
datasets_sub = dict()
|
| 97 |
+
for key in datasets.keys():
|
| 98 |
+
dataset = datasets[key]
|
| 99 |
+
subsample_len = np.floor((1 - dataset_favour_long) * min_len + dataset_favour_long * len(dataset))
|
| 100 |
+
subsample_len = int(subsample_len)
|
| 101 |
+
|
| 102 |
+
indices = torch.randperm(len(dataset))
|
| 103 |
+
subsampled_dataset = Subset(dataset, indices[0:subsample_len])
|
| 104 |
+
datasets_sub[key] = subsampled_dataset
|
| 105 |
+
|
| 106 |
+
incdataset = torch.utils.data.ConcatDataset([datasets_sub[key] for key in datasets_sub.keys()])
|
| 107 |
+
else:
|
| 108 |
+
incdataset = torch.utils.data.ConcatDataset([datasets[key] for key in datasets.keys()])
|
| 109 |
+
|
| 110 |
+
return incdataset
|
external/WildCamera/WildCamera/datasets/MegaDepth.py
ADDED
|
@@ -0,0 +1,76 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, io, glob, natsort, random
|
| 2 |
+
import h5py
|
| 3 |
+
import torch
|
| 4 |
+
import numpy as np
|
| 5 |
+
from PIL import Image
|
| 6 |
+
from torchvision import transforms
|
| 7 |
+
|
| 8 |
+
class MegaDepth:
|
| 9 |
+
def __init__(self, data_root, ht=384, wt=512, shuffleseed=None, split='train') -> None:
|
| 10 |
+
split_path = os.path.join('splits', 'megadepth_{}.txt'.format(split))
|
| 11 |
+
|
| 12 |
+
with open(split_path) as file:
|
| 13 |
+
data_names = [line.rstrip() for line in file]
|
| 14 |
+
|
| 15 |
+
if split == 'train':
|
| 16 |
+
if shuffleseed is not None:
|
| 17 |
+
random.seed(shuffleseed)
|
| 18 |
+
random.shuffle(data_names)
|
| 19 |
+
|
| 20 |
+
self.data_root = data_root
|
| 21 |
+
|
| 22 |
+
self.wt, self.ht = wt, ht
|
| 23 |
+
self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
|
| 24 |
+
self.tensor = transforms.ToTensor()
|
| 25 |
+
self.data_names = data_names
|
| 26 |
+
|
| 27 |
+
self.datasetname = 'MegaDepth'
|
| 28 |
+
|
| 29 |
+
def __len__(self):
|
| 30 |
+
return len(self.data_names)
|
| 31 |
+
|
| 32 |
+
def load_im(self, im_ref):
|
| 33 |
+
im = Image.open(im_ref)
|
| 34 |
+
return im
|
| 35 |
+
|
| 36 |
+
def __getitem__(self, idx):
|
| 37 |
+
# read intrinsics of original size
|
| 38 |
+
scene_name, stem_name_color, stem_name_intrinsic = self.data_names[idx].split(' ')
|
| 39 |
+
h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
|
| 40 |
+
|
| 41 |
+
with h5py.File(h5pypath, 'r') as hf:
|
| 42 |
+
K_color = np.array(hf['intrinsic'][stem_name_intrinsic])
|
| 43 |
+
|
| 44 |
+
# Load positive pair data
|
| 45 |
+
rgb = self.load_im(io.BytesIO(np.array(hf['color'][stem_name_color])))
|
| 46 |
+
w, h = rgb.size
|
| 47 |
+
rgb = rgb.resize((self.wt, self.ht))
|
| 48 |
+
|
| 49 |
+
scaleM = np.eye(3)
|
| 50 |
+
scaleM[0, 0] = self.wt / w
|
| 51 |
+
scaleM[1, 1] = self.ht / h
|
| 52 |
+
aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
|
| 53 |
+
|
| 54 |
+
K = torch.from_numpy(scaleM @ K_color).float()
|
| 55 |
+
|
| 56 |
+
# Recompute camera intrinsic matrix due to the resize
|
| 57 |
+
rgb = self.normalize(self.tensor(rgb))
|
| 58 |
+
|
| 59 |
+
# Save RAW
|
| 60 |
+
K_raw = torch.clone(K)
|
| 61 |
+
rgb_raw = torch.clone(rgb)
|
| 62 |
+
|
| 63 |
+
data_dict = {
|
| 64 |
+
'K': K,
|
| 65 |
+
'rgb': rgb,
|
| 66 |
+
'K_raw': K_raw,
|
| 67 |
+
'rgb_raw': rgb_raw,
|
| 68 |
+
'aspect_ratio_restoration': aspect_ratio_restoration,
|
| 69 |
+
'datasetname': self.datasetname,
|
| 70 |
+
# Only Required in Crop Evaluation
|
| 71 |
+
'T': torch.eye(3, dtype=torch.float32),
|
| 72 |
+
'scaleM': scaleM.astype(np.float32),
|
| 73 |
+
'size_wo_change': (h, w),
|
| 74 |
+
}
|
| 75 |
+
|
| 76 |
+
return data_dict
|
external/WildCamera/WildCamera/evaluation/evaluate_crop.py
ADDED
|
@@ -0,0 +1,157 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import sys, os, time, inspect, tabulate
|
| 2 |
+
import tqdm
|
| 3 |
+
import numpy as np
|
| 4 |
+
|
| 5 |
+
import torch
|
| 6 |
+
import torch.distributed as dist
|
| 7 |
+
from torch.utils.data import DataLoader
|
| 8 |
+
from torchvision import transforms, ops
|
| 9 |
+
|
| 10 |
+
from loguru import logger
|
| 11 |
+
|
| 12 |
+
from tools.tools import to_cuda, DistributedSamplerNoEvenlyDivisible
|
| 13 |
+
from tools.visualization import tensor2rgb
|
| 14 |
+
from tools.calibrator import MonocularCalibrator
|
| 15 |
+
from WildCamera.datasets.IncdDataset import IncdDataset
|
| 16 |
+
|
| 17 |
+
class EvaluateCrop:
|
| 18 |
+
def __init__(self):
|
| 19 |
+
return
|
| 20 |
+
|
| 21 |
+
def evaluate(self, model, args, group=None):
|
| 22 |
+
measurements_all = dict()
|
| 23 |
+
|
| 24 |
+
for dataaset in args.datasets_eval:
|
| 25 |
+
|
| 26 |
+
augmentation = True
|
| 27 |
+
|
| 28 |
+
val_dataset = IncdDataset(
|
| 29 |
+
data_root=args.data_path,
|
| 30 |
+
datasets_included=[dataaset],
|
| 31 |
+
ht=args.input_height,
|
| 32 |
+
wt=args.input_width,
|
| 33 |
+
augmentation=augmentation,
|
| 34 |
+
shuffleseed=None,
|
| 35 |
+
split='test',
|
| 36 |
+
augscale=args.augscale,
|
| 37 |
+
no_change_prob=0.5,
|
| 38 |
+
transformcategory='transform_crop'
|
| 39 |
+
)
|
| 40 |
+
measurements = evaluate_crop(model, val_dataset, args, group)
|
| 41 |
+
measurements_all[dataaset] = measurements
|
| 42 |
+
|
| 43 |
+
if args.gpu == 0:
|
| 44 |
+
errors_all_mean = [['Datasets', 'miou', 'acc']]
|
| 45 |
+
for key in measurements_all.keys():
|
| 46 |
+
measurements = measurements_all[key]
|
| 47 |
+
errors_all_mean.append(
|
| 48 |
+
[
|
| 49 |
+
key,
|
| 50 |
+
measurements['miou'],
|
| 51 |
+
measurements['acc'],
|
| 52 |
+
]
|
| 53 |
+
)
|
| 54 |
+
logger.info(
|
| 55 |
+
"\n" +
|
| 56 |
+
tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
|
| 57 |
+
|
| 58 |
+
torch.cuda.synchronize()
|
| 59 |
+
return
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
def acquire_bbox(K, h, w):
|
| 63 |
+
pts1 = torch.from_numpy(np.array([[0.5, 0.5, 1]])).float().cuda().T
|
| 64 |
+
pts2 = torch.from_numpy(np.array([[w - 0.5, 0.5, 1]])).float().cuda().T
|
| 65 |
+
pts3 = torch.from_numpy(np.array([[w - 0.5, h - 0.5, 1]])).float().cuda().T
|
| 66 |
+
pts4 = torch.from_numpy(np.array([[0.5, h - 0.5, 1]])).float().cuda().T
|
| 67 |
+
|
| 68 |
+
pts1 = K @ pts1
|
| 69 |
+
pts2 = K @ pts2
|
| 70 |
+
pts3 = K @ pts3
|
| 71 |
+
pts4 = K @ pts4
|
| 72 |
+
|
| 73 |
+
ptss = torch.cat([pts1, pts2, pts3, pts4])
|
| 74 |
+
ptss = ptss.squeeze().cpu().numpy()[:, 0:2]
|
| 75 |
+
|
| 76 |
+
ptss_ = [
|
| 77 |
+
min(ptss[:, 0]),
|
| 78 |
+
min(ptss[:, 1]),
|
| 79 |
+
max(ptss[:, 0]),
|
| 80 |
+
max(ptss[:, 1]),
|
| 81 |
+
]
|
| 82 |
+
ptss_ = np.array(ptss_)
|
| 83 |
+
ptss_ = torch.from_numpy(ptss_).float().cuda()
|
| 84 |
+
return ptss_
|
| 85 |
+
|
| 86 |
+
def bbox_auc(errors, thresholds):
|
| 87 |
+
sort_idx = np.argsort(errors)
|
| 88 |
+
errors = np.array(errors.copy())[sort_idx]
|
| 89 |
+
recall = (np.arange(len(errors)) + 1) / len(errors)
|
| 90 |
+
errors = np.r_[0.0, errors]
|
| 91 |
+
recall = np.r_[0.0, recall]
|
| 92 |
+
aucs = []
|
| 93 |
+
for t in thresholds:
|
| 94 |
+
last_index = np.searchsorted(errors, t)
|
| 95 |
+
r = np.r_[recall[:last_index], recall[last_index - 1]]
|
| 96 |
+
e = np.r_[errors[:last_index], t]
|
| 97 |
+
aucs.append(np.trapz(r, x=e) / t)
|
| 98 |
+
return aucs
|
| 99 |
+
|
| 100 |
+
@torch.no_grad()
|
| 101 |
+
def evaluate_crop(model, val_dataset, args, group=None):
|
| 102 |
+
""" In Validation, random sample N Images to mimic a real test set situation """
|
| 103 |
+
monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
|
| 104 |
+
model.eval()
|
| 105 |
+
if group is not None:
|
| 106 |
+
sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
|
| 107 |
+
val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 108 |
+
else:
|
| 109 |
+
val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 110 |
+
|
| 111 |
+
measurements = torch.zeros(len(val_dataset), 2).cuda(device=args.gpu)
|
| 112 |
+
measurements_detection = torch.zeros(len(val_dataset)).cuda(device=args.gpu)
|
| 113 |
+
world_size = args.world_size
|
| 114 |
+
|
| 115 |
+
th = 0.2
|
| 116 |
+
|
| 117 |
+
for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
|
| 118 |
+
sample_batched = to_cuda(data_blob)
|
| 119 |
+
rgb, K, scaleM, T, K_raw, rgb_raw = sample_batched['rgb'], sample_batched['K'], sample_batched['scaleM'], sample_batched['T'], sample_batched['K_raw'], sample_batched['rgb_raw']
|
| 120 |
+
|
| 121 |
+
incidence = model(rgb)
|
| 122 |
+
Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
|
| 123 |
+
|
| 124 |
+
Kest_wo_change = torch.inverse(scaleM) @ Kest
|
| 125 |
+
fx_fy = (Kest_wo_change[0, 0, 0] / Kest_wo_change[0, 1, 1]).item()
|
| 126 |
+
fx_fy = max(fx_fy, 1 / fx_fy)
|
| 127 |
+
fx_fy = abs(fx_fy - 1)
|
| 128 |
+
diffx = abs(Kest_wo_change[0, 0, 2].item() - sample_batched['size_wo_change'][1].item() / 2)
|
| 129 |
+
diffx = diffx / sample_batched['size_wo_change'][1].item()
|
| 130 |
+
diffy = abs(Kest_wo_change[0, 1, 2].item() - sample_batched['size_wo_change'][0].item() / 2)
|
| 131 |
+
diffy = diffy / sample_batched['size_wo_change'][0].item()
|
| 132 |
+
diffb = max(diffx, diffy)
|
| 133 |
+
genuen = (fx_fy < th) and (diffb < th)
|
| 134 |
+
|
| 135 |
+
genuen_gt = ((T[0].cpu() - torch.eye(3)).abs().max() < 1e-2).item()
|
| 136 |
+
|
| 137 |
+
if genuen == genuen_gt:
|
| 138 |
+
measurements_detection[val_id * world_size + args.gpu] = 1
|
| 139 |
+
|
| 140 |
+
b, _, h, w = rgb.shape
|
| 141 |
+
bbox_est = acquire_bbox(K_raw @ torch.inverse(Kest), h, w)
|
| 142 |
+
bbox_gt = acquire_bbox(K_raw @ torch.inverse(Kgt), h, w)
|
| 143 |
+
|
| 144 |
+
iou = ops.box_iou(bbox_gt.unsqueeze(0), bbox_est.unsqueeze(0))
|
| 145 |
+
measurements[val_id * world_size + args.gpu, 0] = iou.squeeze().item()
|
| 146 |
+
measurements[val_id * world_size + args.gpu, 1] = 1 - float(genuen_gt)
|
| 147 |
+
|
| 148 |
+
if group is not None:
|
| 149 |
+
dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
|
| 150 |
+
|
| 151 |
+
measurements = measurements.cpu().numpy()
|
| 152 |
+
miou = np.sum(measurements[:, 0] * measurements[:, 1]) / np.sum(measurements[:, 1])
|
| 153 |
+
measurements = {
|
| 154 |
+
'miou': miou,
|
| 155 |
+
'acc': measurements_detection.mean().item()
|
| 156 |
+
}
|
| 157 |
+
return measurements
|
external/WildCamera/WildCamera/evaluation/evaluate_fov.py
ADDED
|
@@ -0,0 +1,142 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import sys, os, time, inspect, tabulate
|
| 2 |
+
import tqdm
|
| 3 |
+
import numpy as np
|
| 4 |
+
|
| 5 |
+
import torch
|
| 6 |
+
import torch.distributed as dist
|
| 7 |
+
from torch.utils.data import DataLoader
|
| 8 |
+
|
| 9 |
+
from loguru import logger
|
| 10 |
+
|
| 11 |
+
from tools.tools import to_cuda
|
| 12 |
+
from tools.calibrator import MonocularCalibrator
|
| 13 |
+
from tools.tools import DistributedSamplerNoEvenlyDivisible
|
| 14 |
+
from WildCamera.datasets.IncdDataset import IncdDataset
|
| 15 |
+
|
| 16 |
+
class EvaluateFov:
|
| 17 |
+
|
| 18 |
+
def __init__(self):
|
| 19 |
+
self.min_fovy = np.inf
|
| 20 |
+
|
| 21 |
+
def evaluate(self, model, args, steps, writer=None, group=None, wtassumption=False, split='test'):
|
| 22 |
+
|
| 23 |
+
if len(args.datasets_eval) > 1:
|
| 24 |
+
autosave = False
|
| 25 |
+
else:
|
| 26 |
+
autosave = True
|
| 27 |
+
|
| 28 |
+
measurements_all = dict()
|
| 29 |
+
|
| 30 |
+
for dataaset in args.datasets_eval:
|
| 31 |
+
|
| 32 |
+
if dataaset in args.datasets_train:
|
| 33 |
+
augmentation = True
|
| 34 |
+
else:
|
| 35 |
+
augmentation = False
|
| 36 |
+
|
| 37 |
+
val_dataset = IncdDataset(
|
| 38 |
+
data_root=args.data_path,
|
| 39 |
+
datasets_included=[dataaset],
|
| 40 |
+
ht=args.input_height,
|
| 41 |
+
wt=args.input_width,
|
| 42 |
+
augmentation=augmentation,
|
| 43 |
+
shuffleseed=None,
|
| 44 |
+
split=split,
|
| 45 |
+
augscale=args.augscale,
|
| 46 |
+
no_change_prob=0.0
|
| 47 |
+
)
|
| 48 |
+
measurements = evaluate_fov(model, val_dataset, args, group, wtassumption=wtassumption)
|
| 49 |
+
measurements_all[dataaset] = measurements
|
| 50 |
+
|
| 51 |
+
if args.gpu == 0:
|
| 52 |
+
if writer is not None and (steps > 0):
|
| 53 |
+
for k in measurements.keys():
|
| 54 |
+
writer.add_scalar('Eval_{}/{}'.format(dataaset, k), measurements[k], steps)
|
| 55 |
+
|
| 56 |
+
if args.gpu == 0 and autosave and (steps > 0):
|
| 57 |
+
if measurements['error_fovy'] < self.min_fovy:
|
| 58 |
+
self.min_fovy = measurements['error_fovy']
|
| 59 |
+
svpath = os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'min_fovy.ckpt')
|
| 60 |
+
try:
|
| 61 |
+
model_ = model.module
|
| 62 |
+
except:
|
| 63 |
+
model_ = model
|
| 64 |
+
torch.save(model_.state_dict(), svpath)
|
| 65 |
+
logger.info("Save to %s" % svpath)
|
| 66 |
+
|
| 67 |
+
if args.gpu == 0 and (not autosave) and (steps > 0):
|
| 68 |
+
try:
|
| 69 |
+
model_ = model.module
|
| 70 |
+
except:
|
| 71 |
+
model_ = model
|
| 72 |
+
torch.save(
|
| 73 |
+
model_.state_dict(),
|
| 74 |
+
os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'step_{}.ckpt'.format(str(steps)))
|
| 75 |
+
)
|
| 76 |
+
|
| 77 |
+
if args.gpu == 0:
|
| 78 |
+
errors_all_mean = [['Datasets', 'error_fovy', 'error_fovy_medi']]
|
| 79 |
+
for key in measurements_all.keys():
|
| 80 |
+
measurements = measurements_all[key]
|
| 81 |
+
errors_all_mean.append(
|
| 82 |
+
[
|
| 83 |
+
key,
|
| 84 |
+
measurements['error_fovy'],
|
| 85 |
+
measurements['error_fovy_medi'],
|
| 86 |
+
]
|
| 87 |
+
)
|
| 88 |
+
logger.info(
|
| 89 |
+
"\n==================== Performance at Step %d with Assumption %d ====================\n" % (steps, wtassumption) +
|
| 90 |
+
tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
|
| 91 |
+
|
| 92 |
+
torch.cuda.synchronize()
|
| 93 |
+
return
|
| 94 |
+
|
| 95 |
+
@torch.no_grad()
|
| 96 |
+
def evaluate_fov(model, val_dataset, args, group=None, wtassumption=False):
|
| 97 |
+
""" In Validation, random sample N Images to mimic a real test set situation """
|
| 98 |
+
monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
|
| 99 |
+
model.eval()
|
| 100 |
+
if args.world_size > 1:
|
| 101 |
+
sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
|
| 102 |
+
val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 103 |
+
else:
|
| 104 |
+
val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 105 |
+
|
| 106 |
+
measurements = torch.zeros(len(val_dataset), 6).cuda(device=args.gpu)
|
| 107 |
+
world_size = args.world_size
|
| 108 |
+
|
| 109 |
+
for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
|
| 110 |
+
sample_batched = to_cuda(data_blob)
|
| 111 |
+
rgb, K, r = sample_batched['rgb'], sample_batched['K'], sample_batched['aspect_ratio_restoration']
|
| 112 |
+
|
| 113 |
+
incidence = model(rgb)
|
| 114 |
+
if not wtassumption:
|
| 115 |
+
Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
|
| 116 |
+
else:
|
| 117 |
+
Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_1DoF(incidence, r=r.item())
|
| 118 |
+
|
| 119 |
+
b, _, h, w = rgb.shape
|
| 120 |
+
device = rgb.device
|
| 121 |
+
|
| 122 |
+
fovy_est, fovy_gt = 2 * torch.arctan(h / Kest[1, 1] / 2), sample_batched['fovy']
|
| 123 |
+
error_fovy = np.abs(np.degrees(fovy_est.item()) - np.degrees(fovy_gt.item()))
|
| 124 |
+
|
| 125 |
+
error_all = np.array([error_fovy])
|
| 126 |
+
error_all = torch.from_numpy(error_all).float().to(device)
|
| 127 |
+
|
| 128 |
+
measurements[val_id * world_size + args.gpu] += error_all
|
| 129 |
+
|
| 130 |
+
if args.world_size > 1:
|
| 131 |
+
dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
|
| 132 |
+
|
| 133 |
+
zero_entry = torch.sum(measurements.abs(), dim=1) == 0
|
| 134 |
+
assert torch.sum(zero_entry) == 0
|
| 135 |
+
|
| 136 |
+
fovy_median = torch.median(measurements.squeeze())
|
| 137 |
+
measurements = torch.mean(measurements, dim=0)
|
| 138 |
+
measurements = {
|
| 139 |
+
'error_fovy': measurements[0].item(),
|
| 140 |
+
'error_fovy_medi': fovy_median.item()
|
| 141 |
+
}
|
| 142 |
+
return measurements
|
external/WildCamera/WildCamera/evaluation/evaluate_intrinsic.py
ADDED
|
@@ -0,0 +1,144 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import sys, os, time, inspect, tabulate
|
| 2 |
+
import tqdm
|
| 3 |
+
import numpy as np
|
| 4 |
+
|
| 5 |
+
import torch
|
| 6 |
+
import torch.distributed as dist
|
| 7 |
+
from torch.utils.data import DataLoader
|
| 8 |
+
|
| 9 |
+
from loguru import logger
|
| 10 |
+
|
| 11 |
+
from tools.tools import to_cuda
|
| 12 |
+
from tools.calibrator import MonocularCalibrator
|
| 13 |
+
from tools.evaluation import compute_intrinsic_measure
|
| 14 |
+
from tools.tools import DistributedSamplerNoEvenlyDivisible
|
| 15 |
+
from WildCamera.datasets.IncdDataset import IncdDataset
|
| 16 |
+
|
| 17 |
+
class EvaluateIntrinsic:
|
| 18 |
+
def __init__(self):
|
| 19 |
+
self.min_focal = 1e10
|
| 20 |
+
|
| 21 |
+
def evaluate(self, model, args, steps, writer=None, group=None, wtassumption=False):
|
| 22 |
+
|
| 23 |
+
if len(args.datasets_eval) > 1:
|
| 24 |
+
autosave = False
|
| 25 |
+
else:
|
| 26 |
+
autosave = True
|
| 27 |
+
|
| 28 |
+
measurements_all = dict()
|
| 29 |
+
|
| 30 |
+
for dataaset in args.datasets_eval:
|
| 31 |
+
|
| 32 |
+
if dataaset in args.datasets_train:
|
| 33 |
+
augmentation = True
|
| 34 |
+
else:
|
| 35 |
+
augmentation = False
|
| 36 |
+
|
| 37 |
+
val_dataset = IncdDataset(
|
| 38 |
+
data_root=args.data_path,
|
| 39 |
+
datasets_included=[dataaset],
|
| 40 |
+
ht=args.input_height,
|
| 41 |
+
wt=args.input_width,
|
| 42 |
+
augmentation=augmentation,
|
| 43 |
+
shuffleseed=None,
|
| 44 |
+
split='test',
|
| 45 |
+
augscale=args.augscale,
|
| 46 |
+
no_change_prob=0.0,
|
| 47 |
+
transformcategory='transform_calibration'
|
| 48 |
+
)
|
| 49 |
+
measurements = evaluate_intrinsic(model, val_dataset, args, group, wtassumption=wtassumption)
|
| 50 |
+
measurements_all[dataaset] = measurements
|
| 51 |
+
|
| 52 |
+
if args.gpu == 0:
|
| 53 |
+
if writer is not None and (steps > 0):
|
| 54 |
+
for k in measurements.keys():
|
| 55 |
+
writer.add_scalar('Eval_{}/{}'.format(dataaset, k), measurements[k], steps)
|
| 56 |
+
|
| 57 |
+
if args.gpu == 0 and autosave and (steps > 0):
|
| 58 |
+
if measurements['error_f'] < self.min_focal:
|
| 59 |
+
self.min_focal = measurements['error_f']
|
| 60 |
+
svpath = os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'min_focal.ckpt')
|
| 61 |
+
torch.save(model.state_dict(), svpath)
|
| 62 |
+
logger.info("Save to %s" % svpath)
|
| 63 |
+
|
| 64 |
+
if args.gpu == 0 and (not autosave) and (steps > 0):
|
| 65 |
+
torch.save(model.state_dict(), os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'step_{}.ckpt'.format(str(steps))))
|
| 66 |
+
|
| 67 |
+
if args.gpu == 0:
|
| 68 |
+
errors_all_mean = [['Datasets', 'error_fx', 'error_fy', 'error_f', 'error_bx', 'error_by', 'error_b']]
|
| 69 |
+
for key in measurements_all.keys():
|
| 70 |
+
measurements = measurements_all[key]
|
| 71 |
+
errors_all_mean.append(
|
| 72 |
+
[
|
| 73 |
+
key,
|
| 74 |
+
measurements['error_fx'],
|
| 75 |
+
measurements['error_fy'],
|
| 76 |
+
measurements['error_f'],
|
| 77 |
+
measurements['error_bx'],
|
| 78 |
+
measurements['error_by'],
|
| 79 |
+
measurements['error_b']
|
| 80 |
+
]
|
| 81 |
+
)
|
| 82 |
+
logger.info(
|
| 83 |
+
"\n==================== Performance at Step %d with Assumption %d ====================\n" % (steps, wtassumption) +
|
| 84 |
+
tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
|
| 85 |
+
|
| 86 |
+
torch.cuda.synchronize()
|
| 87 |
+
return
|
| 88 |
+
|
| 89 |
+
@torch.no_grad()
|
| 90 |
+
def evaluate_intrinsic(model, val_dataset, args, group=None, wtassumption=False):
|
| 91 |
+
""" In Validation, random sample N Images to mimic a real test set situation """
|
| 92 |
+
monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
|
| 93 |
+
model.eval()
|
| 94 |
+
if args.world_size > 1:
|
| 95 |
+
sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
|
| 96 |
+
val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 97 |
+
else:
|
| 98 |
+
val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
|
| 99 |
+
|
| 100 |
+
measurements = torch.zeros(len(val_dataset), 6).cuda(device=args.gpu)
|
| 101 |
+
world_size = args.world_size
|
| 102 |
+
|
| 103 |
+
for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
|
| 104 |
+
sample_batched = to_cuda(data_blob)
|
| 105 |
+
rgb, K, r = sample_batched['rgb'], sample_batched['K'], sample_batched['aspect_ratio_restoration']
|
| 106 |
+
|
| 107 |
+
incidence = model(rgb)
|
| 108 |
+
if not wtassumption:
|
| 109 |
+
Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
|
| 110 |
+
else:
|
| 111 |
+
Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_1DoF(incidence, r=r.item())
|
| 112 |
+
|
| 113 |
+
b, _, h, w = rgb.shape
|
| 114 |
+
device = rgb.device
|
| 115 |
+
|
| 116 |
+
error_fx, error_fy, error_f, error_bx, error_by, error_b = compute_intrinsic_measure(
|
| 117 |
+
Kest=torch.clone(Kest),
|
| 118 |
+
Kgt=torch.clone(Kgt),
|
| 119 |
+
h=h,
|
| 120 |
+
w=w
|
| 121 |
+
)
|
| 122 |
+
|
| 123 |
+
error_all = np.array([error_fx, error_fy, error_f, error_bx, error_by, error_b])
|
| 124 |
+
error_all = torch.from_numpy(error_all).float().to(device)
|
| 125 |
+
|
| 126 |
+
measurements[val_id * world_size + args.gpu] += error_all
|
| 127 |
+
|
| 128 |
+
if args.world_size > 1:
|
| 129 |
+
dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
|
| 130 |
+
|
| 131 |
+
zero_entry = torch.sum(measurements.abs(), dim=1) == 0
|
| 132 |
+
assert torch.sum(zero_entry) == 0
|
| 133 |
+
|
| 134 |
+
measurements = torch.mean(measurements, dim=0)
|
| 135 |
+
measurements = {
|
| 136 |
+
'error_fx': measurements[0].item(),
|
| 137 |
+
'error_fy': measurements[1].item(),
|
| 138 |
+
'error_f': measurements[2].item(),
|
| 139 |
+
'error_bx': measurements[3].item(),
|
| 140 |
+
'error_by': measurements[4].item(),
|
| 141 |
+
'error_b': measurements[5].item()
|
| 142 |
+
}
|
| 143 |
+
|
| 144 |
+
return measurements
|
external/WildCamera/WildCamera/evaluation/evaluate_pose.py
ADDED
|
@@ -0,0 +1,209 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Put most Common Functions Here
|
| 2 |
+
import cv2
|
| 3 |
+
import torch
|
| 4 |
+
import numpy as np
|
| 5 |
+
import torch.nn.functional as F
|
| 6 |
+
|
| 7 |
+
from einops.einops import rearrange
|
| 8 |
+
|
| 9 |
+
# -- # Pose Benchmark
|
| 10 |
+
def angle_error_mat(R1, R2):
|
| 11 |
+
cos = (np.trace(np.dot(R1.T, R2)) - 1) / 2
|
| 12 |
+
cos = np.clip(cos, -1.0, 1.0) # numercial errors can make it out of bounds
|
| 13 |
+
return np.rad2deg(np.abs(np.arccos(cos)))
|
| 14 |
+
|
| 15 |
+
def angle_error_vec(v1, v2):
|
| 16 |
+
n = np.linalg.norm(v1) * np.linalg.norm(v2)
|
| 17 |
+
return np.rad2deg(np.arccos(np.clip(np.dot(v1, v2) / n, -1.0, 1.0)))
|
| 18 |
+
|
| 19 |
+
def compute_pose_error(T_0to1, R, t):
|
| 20 |
+
R_gt = T_0to1[:3, :3]
|
| 21 |
+
t_gt = T_0to1[:3, 3]
|
| 22 |
+
error_t = angle_error_vec(t.squeeze(), t_gt)
|
| 23 |
+
error_t = np.minimum(error_t, 180 - error_t) # ambiguity of E estimation
|
| 24 |
+
error_R = angle_error_mat(R, R_gt)
|
| 25 |
+
return error_t, error_R
|
| 26 |
+
|
| 27 |
+
def pose_auc(errors, thresholds):
|
| 28 |
+
sort_idx = np.argsort(errors)
|
| 29 |
+
errors = np.array(errors.copy())[sort_idx]
|
| 30 |
+
recall = (np.arange(len(errors)) + 1) / len(errors)
|
| 31 |
+
errors = np.r_[0.0, errors]
|
| 32 |
+
recall = np.r_[0.0, recall]
|
| 33 |
+
aucs = []
|
| 34 |
+
for t in thresholds:
|
| 35 |
+
last_index = np.searchsorted(errors, t)
|
| 36 |
+
r = np.r_[recall[:last_index], recall[last_index - 1]]
|
| 37 |
+
e = np.r_[errors[:last_index], t]
|
| 38 |
+
aucs.append(np.trapz(r, x=e) / t)
|
| 39 |
+
return aucs
|
| 40 |
+
|
| 41 |
+
def compute_relative_pose(R1, t1, R2, t2):
|
| 42 |
+
rots = R2 @ (R1.T)
|
| 43 |
+
trans = -rots @ t1 + t2
|
| 44 |
+
return rots, trans
|
| 45 |
+
|
| 46 |
+
def rotate_intrinsic(K, n):
|
| 47 |
+
base_rot = np.array([[0, 1, 0], [-1, 0, 0], [0, 0, 1]])
|
| 48 |
+
rot = np.linalg.matrix_power(base_rot, n)
|
| 49 |
+
return rot @ K
|
| 50 |
+
|
| 51 |
+
def estimate_pose(kpts0, kpts1, K0, K1, norm_thresh, conf=0.99999):
|
| 52 |
+
if len(kpts0) < 5:
|
| 53 |
+
return None
|
| 54 |
+
K0inv = np.linalg.inv(K0[:2,:2])
|
| 55 |
+
K1inv = np.linalg.inv(K1[:2,:2])
|
| 56 |
+
|
| 57 |
+
kpts0 = (K0inv @ (kpts0-K0[None,:2,2]).T).T
|
| 58 |
+
kpts1 = (K1inv @ (kpts1-K1[None,:2,2]).T).T
|
| 59 |
+
|
| 60 |
+
E, mask = cv2.findEssentialMat(
|
| 61 |
+
kpts0, kpts1, np.eye(3), threshold=norm_thresh, prob=conf, method=cv2.RANSAC
|
| 62 |
+
)
|
| 63 |
+
|
| 64 |
+
ret = None
|
| 65 |
+
if E is not None:
|
| 66 |
+
best_num_inliers = 0
|
| 67 |
+
|
| 68 |
+
for _E in np.split(E, len(E) / 3):
|
| 69 |
+
n, R, t, _ = cv2.recoverPose(_E, kpts0, kpts1, np.eye(3), 1e9, mask=mask)
|
| 70 |
+
if n > best_num_inliers:
|
| 71 |
+
best_num_inliers = n
|
| 72 |
+
ret = (R, t, mask.ravel() > 0)
|
| 73 |
+
return ret
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
# -- # Compute Error in EPE
|
| 77 |
+
|
| 78 |
+
def compute_flow_metrics(depth1, depth2, T_1to2, K1, K2, dense_matches):
|
| 79 |
+
"""[summary]
|
| 80 |
+
|
| 81 |
+
Args:
|
| 82 |
+
H ([type]): [description]
|
| 83 |
+
scale ([type]): [description]
|
| 84 |
+
|
| 85 |
+
Returns:
|
| 86 |
+
[type]: [description]
|
| 87 |
+
"""
|
| 88 |
+
dense_matches = rearrange(dense_matches, "b d h w -> b h w d")
|
| 89 |
+
b, h1, w1, d = dense_matches.shape
|
| 90 |
+
x1_n = torch.meshgrid(
|
| 91 |
+
*[
|
| 92 |
+
torch.linspace(
|
| 93 |
+
-1 + 1 / n, 1 - 1 / n, n, device=dense_matches.device
|
| 94 |
+
)
|
| 95 |
+
for n in (b, h1, w1)
|
| 96 |
+
]
|
| 97 |
+
)
|
| 98 |
+
x1_n = torch.stack((x1_n[2], x1_n[1]), dim=-1).reshape(b, h1 * w1, 2)
|
| 99 |
+
mask, x2 = warp_kpts(
|
| 100 |
+
x1_n.double(),
|
| 101 |
+
depth1.double(),
|
| 102 |
+
depth2.double(),
|
| 103 |
+
T_1to2.double(),
|
| 104 |
+
K1.double(),
|
| 105 |
+
K2.double(),
|
| 106 |
+
)
|
| 107 |
+
mask_validation = mask.float().reshape(b, h1, w1) > 0
|
| 108 |
+
x2 = x2.reshape(b, h1, w1, 2)
|
| 109 |
+
|
| 110 |
+
dense_matches = unnorm_coords_Numpystyle(
|
| 111 |
+
rearrange(dense_matches, "b h w d -> b d h w"),
|
| 112 |
+
h=h1, w=w1
|
| 113 |
+
)
|
| 114 |
+
|
| 115 |
+
x2 = unnorm_coords_Numpystyle(
|
| 116 |
+
rearrange(x2, "b h w d -> b d h w"),
|
| 117 |
+
h=h1, w=w1
|
| 118 |
+
)
|
| 119 |
+
|
| 120 |
+
epe = (dense_matches - x2).norm(dim=1) # *scale?
|
| 121 |
+
|
| 122 |
+
evaluated_pixel_num = torch.sum(mask_validation)
|
| 123 |
+
epe_sum = torch.sum(epe[mask_validation])
|
| 124 |
+
px1 = torch.sum((epe < 1)[mask_validation])
|
| 125 |
+
px5 = torch.sum((epe < 5)[mask_validation])
|
| 126 |
+
px8 = torch.sum((epe < 8)[mask_validation])
|
| 127 |
+
|
| 128 |
+
sub_measurements = torch.stack([epe_sum, px1, px5, px8, evaluated_pixel_num])
|
| 129 |
+
return sub_measurements
|
| 130 |
+
|
| 131 |
+
|
| 132 |
+
|
| 133 |
+
@torch.no_grad()
|
| 134 |
+
def warp_kpts(kpts0, depth0, depth1, T_0to1, K0, K1):
|
| 135 |
+
"""Warp kpts0 from I0 to I1 with depth, K and Rt
|
| 136 |
+
Also check covisibility and depth consistency.
|
| 137 |
+
Depth is consistent if relative error < 0.2 (hard-coded).
|
| 138 |
+
# https://github.com/zju3dv/LoFTR/blob/94e98b695be18acb43d5d3250f52226a8e36f839/src/loftr/utils/geometry.py adapted from here
|
| 139 |
+
Args:
|
| 140 |
+
kpts0 (torch.Tensor): [N, L, 2] - <x, y>, should be normalized in (-1,1)
|
| 141 |
+
depth0 (torch.Tensor): [N, H, W],
|
| 142 |
+
depth1 (torch.Tensor): [N, H, W],
|
| 143 |
+
T_0to1 (torch.Tensor): [N, 3, 4],
|
| 144 |
+
K0 (torch.Tensor): [N, 3, 3],
|
| 145 |
+
K1 (torch.Tensor): [N, 3, 3],
|
| 146 |
+
Returns:
|
| 147 |
+
calculable_mask (torch.Tensor): [N, L]
|
| 148 |
+
warped_keypoints0 (torch.Tensor): [N, L, 2] <x0_hat, y1_hat>
|
| 149 |
+
"""
|
| 150 |
+
(
|
| 151 |
+
n,
|
| 152 |
+
h,
|
| 153 |
+
w,
|
| 154 |
+
) = depth0.shape
|
| 155 |
+
kpts0_depth = F.grid_sample(depth0[:, None], kpts0[:, :, None], mode="bilinear", align_corners=False)[
|
| 156 |
+
:, 0, :, 0
|
| 157 |
+
]
|
| 158 |
+
kpts0 = torch.stack(
|
| 159 |
+
(w * (kpts0[..., 0] + 1) / 2, h * (kpts0[..., 1] + 1) / 2), dim=-1
|
| 160 |
+
) # [-1+1/h, 1-1/h] -> [0.5, h-0.5]
|
| 161 |
+
# Sample depth, get calculable_mask on depth != 0
|
| 162 |
+
nonzero_mask = kpts0_depth != 0
|
| 163 |
+
|
| 164 |
+
# Unproject
|
| 165 |
+
kpts0_h = (
|
| 166 |
+
torch.cat([kpts0, torch.ones_like(kpts0[:, :, [0]])], dim=-1)
|
| 167 |
+
* kpts0_depth[..., None]
|
| 168 |
+
) # (N, L, 3)
|
| 169 |
+
kpts0_n = K0.inverse() @ kpts0_h.transpose(2, 1) # (N, 3, L)
|
| 170 |
+
kpts0_cam = kpts0_n
|
| 171 |
+
|
| 172 |
+
# Rigid Transform
|
| 173 |
+
w_kpts0_cam = T_0to1[:, :3, :3] @ kpts0_cam + T_0to1[:, :3, [3]] # (N, 3, L)
|
| 174 |
+
w_kpts0_depth_computed = w_kpts0_cam[:, 2, :]
|
| 175 |
+
|
| 176 |
+
# Project
|
| 177 |
+
w_kpts0_h = (K1 @ w_kpts0_cam).transpose(2, 1) # (N, L, 3)
|
| 178 |
+
w_kpts0 = w_kpts0_h[:, :, :2] / (
|
| 179 |
+
w_kpts0_h[:, :, [2]] + 1e-4
|
| 180 |
+
) # (N, L, 2), +1e-4 to avoid zero depth
|
| 181 |
+
|
| 182 |
+
# Covisible Check
|
| 183 |
+
h, w = depth1.shape[1:3]
|
| 184 |
+
covisible_mask = (
|
| 185 |
+
(w_kpts0[:, :, 0] > 0)
|
| 186 |
+
* (w_kpts0[:, :, 0] < w - 1)
|
| 187 |
+
* (w_kpts0[:, :, 1] > 0)
|
| 188 |
+
* (w_kpts0[:, :, 1] < h - 1)
|
| 189 |
+
)
|
| 190 |
+
w_kpts0 = torch.stack(
|
| 191 |
+
(2 * w_kpts0[..., 0] / w - 1, 2 * w_kpts0[..., 1] / h - 1), dim=-1
|
| 192 |
+
) # from [0.5,h-0.5] -> [-1+1/h, 1-1/h]
|
| 193 |
+
# w_kpts0[~covisible_mask, :] = -5 # xd
|
| 194 |
+
|
| 195 |
+
w_kpts0_depth = F.grid_sample(
|
| 196 |
+
depth1[:, None], w_kpts0[:, :, None], mode="bilinear", align_corners=False
|
| 197 |
+
)[:, 0, :, 0]
|
| 198 |
+
consistent_mask = (
|
| 199 |
+
(w_kpts0_depth - w_kpts0_depth_computed) / w_kpts0_depth
|
| 200 |
+
).abs() < 0.05
|
| 201 |
+
valid_mask = nonzero_mask * covisible_mask * consistent_mask
|
| 202 |
+
|
| 203 |
+
return valid_mask, w_kpts0
|
| 204 |
+
|
| 205 |
+
def unnorm_coords_Numpystyle(coords1, h, w):
|
| 206 |
+
coords1x, coords1y = torch.split(coords1, 1, dim=1)
|
| 207 |
+
coords1x = (coords1x + 1) / 2 * w
|
| 208 |
+
coords1y = (coords1y + 1) / 2 * h
|
| 209 |
+
return torch.cat([coords1x, coords1y], dim=1)
|
external/WildCamera/WildCamera/newcrfs/__init__.py
ADDED
|
File without changes
|
external/WildCamera/WildCamera/newcrfs/__pycache__/swin_transformer.cpython-310.pyc
ADDED
|
Binary file (19.4 kB). View file
|
|
|
external/WildCamera/WildCamera/newcrfs/__pycache__/uper_crf_head.cpython-310.pyc
ADDED
|
Binary file (11.8 kB). View file
|
|
|
external/WildCamera/WildCamera/newcrfs/newcrf_incidencefield.py
ADDED
|
@@ -0,0 +1,196 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
import torch.nn.functional as F
|
| 4 |
+
import PIL.Image as Image
|
| 5 |
+
|
| 6 |
+
import numpy as np
|
| 7 |
+
|
| 8 |
+
from .swin_transformer import SwinTransformer
|
| 9 |
+
from .newcrf_layers import NewCRF
|
| 10 |
+
from .uper_crf_head import PSP
|
| 11 |
+
|
| 12 |
+
from torchvision import transforms
|
| 13 |
+
from tools.calibrator import MonocularCalibrator
|
| 14 |
+
########################################################################################################################
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
class NEWCRFIF(nn.Module):
|
| 18 |
+
"""
|
| 19 |
+
Depth network based on neural window FC-CRFs architecture.
|
| 20 |
+
"""
|
| 21 |
+
def __init__(self, pretrained=None, frozen_stages=-1, version='large07'):
|
| 22 |
+
super().__init__()
|
| 23 |
+
|
| 24 |
+
self.with_auxiliary_head = False
|
| 25 |
+
self.with_neck = False
|
| 26 |
+
|
| 27 |
+
norm_cfg = dict(type='BN', requires_grad=True)
|
| 28 |
+
|
| 29 |
+
window_size = int(version[-2:])
|
| 30 |
+
|
| 31 |
+
if version[:-2] == 'base':
|
| 32 |
+
embed_dim = 128
|
| 33 |
+
depths = [2, 2, 18, 2]
|
| 34 |
+
num_heads = [4, 8, 16, 32]
|
| 35 |
+
in_channels = [128, 256, 512, 1024]
|
| 36 |
+
elif version[:-2] == 'large':
|
| 37 |
+
embed_dim = 192
|
| 38 |
+
depths = [2, 2, 18, 2]
|
| 39 |
+
num_heads = [6, 12, 24, 48]
|
| 40 |
+
in_channels = [192, 384, 768, 1536]
|
| 41 |
+
elif version[:-2] == 'tiny':
|
| 42 |
+
embed_dim = 96
|
| 43 |
+
depths = [2, 2, 6, 2]
|
| 44 |
+
num_heads = [3, 6, 12, 24]
|
| 45 |
+
in_channels = [96, 192, 384, 768]
|
| 46 |
+
|
| 47 |
+
backbone_cfg = dict(
|
| 48 |
+
embed_dim=embed_dim,
|
| 49 |
+
depths=depths,
|
| 50 |
+
num_heads=num_heads,
|
| 51 |
+
window_size=window_size,
|
| 52 |
+
ape=False,
|
| 53 |
+
drop_path_rate=0.3,
|
| 54 |
+
patch_norm=True,
|
| 55 |
+
use_checkpoint=False,
|
| 56 |
+
frozen_stages=frozen_stages
|
| 57 |
+
)
|
| 58 |
+
|
| 59 |
+
embed_dim = 512
|
| 60 |
+
decoder_cfg = dict(
|
| 61 |
+
in_channels=in_channels,
|
| 62 |
+
in_index=[0, 1, 2, 3],
|
| 63 |
+
pool_scales=(1, 2, 3, 6),
|
| 64 |
+
channels=embed_dim,
|
| 65 |
+
dropout_ratio=0.0,
|
| 66 |
+
num_classes=32,
|
| 67 |
+
norm_cfg=norm_cfg,
|
| 68 |
+
align_corners=False
|
| 69 |
+
)
|
| 70 |
+
|
| 71 |
+
self.backbone = SwinTransformer(**backbone_cfg)
|
| 72 |
+
v_dim = decoder_cfg['num_classes']*4
|
| 73 |
+
win = 7
|
| 74 |
+
crf_dims = [128, 256, 512, 1024]
|
| 75 |
+
v_dims = [64, 128, 256, embed_dim]
|
| 76 |
+
self.crf3 = NewCRF(input_dim=in_channels[3], embed_dim=crf_dims[3], window_size=win, v_dim=v_dims[3], num_heads=32)
|
| 77 |
+
self.crf2 = NewCRF(input_dim=in_channels[2], embed_dim=crf_dims[2], window_size=win, v_dim=v_dims[2], num_heads=16)
|
| 78 |
+
self.crf1 = NewCRF(input_dim=in_channels[1], embed_dim=crf_dims[1], window_size=win, v_dim=v_dims[1], num_heads=8)
|
| 79 |
+
self.crf0 = NewCRF(input_dim=in_channels[0], embed_dim=crf_dims[0], window_size=win, v_dim=v_dims[0], num_heads=4)
|
| 80 |
+
|
| 81 |
+
self.decoder = PSP(**decoder_cfg)
|
| 82 |
+
|
| 83 |
+
self.incidence_head = IncidenceHead(input_dim=crf_dims[0], output_dim=3)
|
| 84 |
+
|
| 85 |
+
self.up_mode = 'bilinear'
|
| 86 |
+
if self.up_mode == 'mask':
|
| 87 |
+
self.mask_head = nn.Sequential(
|
| 88 |
+
nn.Conv2d(crf_dims[0], 64, 3, padding=1),
|
| 89 |
+
nn.ReLU(inplace=True),
|
| 90 |
+
nn.Conv2d(64, 16*9, 1, padding=0))
|
| 91 |
+
|
| 92 |
+
self.init_weights(pretrained=pretrained)
|
| 93 |
+
|
| 94 |
+
def init_weights(self, pretrained=None):
|
| 95 |
+
"""Initialize the weights in backbone and heads.
|
| 96 |
+
|
| 97 |
+
Args:
|
| 98 |
+
pretrained (str, optional): Path to pre-trained weights.
|
| 99 |
+
Defaults to None.
|
| 100 |
+
"""
|
| 101 |
+
print(f'== Load encoder backbone from: {pretrained}')
|
| 102 |
+
self.backbone.init_weights(pretrained=pretrained)
|
| 103 |
+
self.decoder.init_weights()
|
| 104 |
+
if self.with_auxiliary_head:
|
| 105 |
+
if isinstance(self.auxiliary_head, nn.ModuleList):
|
| 106 |
+
for aux_head in self.auxiliary_head:
|
| 107 |
+
aux_head.init_weights()
|
| 108 |
+
else:
|
| 109 |
+
self.auxiliary_head.init_weights()
|
| 110 |
+
|
| 111 |
+
def upsample_mask(self, disp, mask):
|
| 112 |
+
""" Upsample disp [H/4, W/4, 1] -> [H, W, 1] using convex combination """
|
| 113 |
+
N, _, H, W = disp.shape
|
| 114 |
+
mask = mask.view(N, 1, 9, 4, 4, H, W)
|
| 115 |
+
mask = torch.softmax(mask, dim=2)
|
| 116 |
+
|
| 117 |
+
up_disp = F.unfold(disp, kernel_size=3, padding=1)
|
| 118 |
+
up_disp = up_disp.view(N, 1, 9, 1, 1, H, W)
|
| 119 |
+
|
| 120 |
+
up_disp = torch.sum(mask * up_disp, dim=2)
|
| 121 |
+
up_disp = up_disp.permute(0, 1, 4, 2, 5, 3)
|
| 122 |
+
return up_disp.reshape(N, 1, 4*H, 4*W)
|
| 123 |
+
|
| 124 |
+
def forward(self, imgs):
|
| 125 |
+
feats = self.backbone(imgs)
|
| 126 |
+
if self.with_neck:
|
| 127 |
+
feats = self.neck(feats)
|
| 128 |
+
|
| 129 |
+
ppm_out = self.decoder(feats)
|
| 130 |
+
|
| 131 |
+
e3 = self.crf3(feats[3], ppm_out)
|
| 132 |
+
e3 = nn.PixelShuffle(2)(e3)
|
| 133 |
+
e2 = self.crf2(feats[2], e3)
|
| 134 |
+
e2 = nn.PixelShuffle(2)(e2)
|
| 135 |
+
e1 = self.crf1(feats[1], e2)
|
| 136 |
+
e1 = nn.PixelShuffle(2)(e1)
|
| 137 |
+
e0 = self.crf0(feats[0], e1)
|
| 138 |
+
|
| 139 |
+
incidence = self.incidence_head(e0, 4)
|
| 140 |
+
|
| 141 |
+
return incidence
|
| 142 |
+
|
| 143 |
+
@torch.no_grad()
|
| 144 |
+
def inference(self, rgb, wtassumption=False):
|
| 145 |
+
self.eval()
|
| 146 |
+
input_wt, input_ht = 640, 480
|
| 147 |
+
w, h = rgb.size
|
| 148 |
+
|
| 149 |
+
rgb = rgb.resize((input_wt, input_ht), Image.Resampling.BILINEAR)
|
| 150 |
+
|
| 151 |
+
scaleM = np.eye(3)
|
| 152 |
+
scaleM[0, 0] = input_wt / w
|
| 153 |
+
scaleM[1, 1] = input_ht / h
|
| 154 |
+
scaleM = torch.from_numpy(scaleM).float().to(next(self.parameters()).device)
|
| 155 |
+
|
| 156 |
+
normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
|
| 157 |
+
totensor = transforms.ToTensor()
|
| 158 |
+
|
| 159 |
+
rgb = normalize(totensor(rgb))
|
| 160 |
+
rgb = rgb.unsqueeze(0).to(next(self.parameters()).device)
|
| 161 |
+
|
| 162 |
+
monocalibrator = MonocularCalibrator()
|
| 163 |
+
incidence = self.forward(rgb)
|
| 164 |
+
|
| 165 |
+
if not wtassumption:
|
| 166 |
+
Kest = monocalibrator.calibrate_camera_4DoF(incidence)
|
| 167 |
+
else:
|
| 168 |
+
r = (scaleM[1, 1] / scaleM[0, 0]).item()
|
| 169 |
+
Kest = monocalibrator.calibrate_camera_1DoF(incidence, r=r)
|
| 170 |
+
|
| 171 |
+
Kest = torch.inverse(scaleM) @ Kest
|
| 172 |
+
return Kest.detach().squeeze().cpu().numpy(), incidence
|
| 173 |
+
|
| 174 |
+
@torch.no_grad()
|
| 175 |
+
def restore_image(self, rgb, intrinsic, fixcrop=True):
|
| 176 |
+
monocalibrator = MonocularCalibrator()
|
| 177 |
+
return monocalibrator.restore_image(rgb, intrinsic, fixcrop)
|
| 178 |
+
|
| 179 |
+
class IncidenceHead(nn.Module):
|
| 180 |
+
def __init__(self, input_dim=100, output_dim=3):
|
| 181 |
+
super(IncidenceHead, self).__init__()
|
| 182 |
+
self.conv1 = nn.Conv2d(input_dim, output_dim, 3, padding=1)
|
| 183 |
+
|
| 184 |
+
def forward(self, x, scale):
|
| 185 |
+
x = self.conv1(x)
|
| 186 |
+
if scale > 1:
|
| 187 |
+
x = upsample(x, scale_factor=scale)
|
| 188 |
+
|
| 189 |
+
incidence = torch.nn.functional.normalize(x, dim=1)
|
| 190 |
+
return incidence
|
| 191 |
+
|
| 192 |
+
|
| 193 |
+
def upsample(x, scale_factor=2, mode="bilinear", align_corners=False):
|
| 194 |
+
"""Upsample input tensor by a factor of 2
|
| 195 |
+
"""
|
| 196 |
+
return F.interpolate(x, scale_factor=scale_factor, mode=mode, align_corners=align_corners)
|
external/WildCamera/WildCamera/newcrfs/newcrf_layers.py
ADDED
|
@@ -0,0 +1,433 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
import torch.nn.functional as F
|
| 4 |
+
import torch.utils.checkpoint as checkpoint
|
| 5 |
+
import numpy as np
|
| 6 |
+
from timm.models.layers import DropPath, to_2tuple, trunc_normal_
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
class Mlp(nn.Module):
|
| 10 |
+
""" Multilayer perceptron."""
|
| 11 |
+
|
| 12 |
+
def __init__(self, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, drop=0.):
|
| 13 |
+
super().__init__()
|
| 14 |
+
out_features = out_features or in_features
|
| 15 |
+
hidden_features = hidden_features or in_features
|
| 16 |
+
self.fc1 = nn.Linear(in_features, hidden_features)
|
| 17 |
+
self.act = act_layer()
|
| 18 |
+
self.fc2 = nn.Linear(hidden_features, out_features)
|
| 19 |
+
self.drop = nn.Dropout(drop)
|
| 20 |
+
|
| 21 |
+
def forward(self, x):
|
| 22 |
+
x = self.fc1(x)
|
| 23 |
+
x = self.act(x)
|
| 24 |
+
x = self.drop(x)
|
| 25 |
+
x = self.fc2(x)
|
| 26 |
+
x = self.drop(x)
|
| 27 |
+
return x
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
def window_partition(x, window_size):
|
| 31 |
+
"""
|
| 32 |
+
Args:
|
| 33 |
+
x: (B, H, W, C)
|
| 34 |
+
window_size (int): window size
|
| 35 |
+
|
| 36 |
+
Returns:
|
| 37 |
+
windows: (num_windows*B, window_size, window_size, C)
|
| 38 |
+
"""
|
| 39 |
+
B, H, W, C = x.shape
|
| 40 |
+
x = x.view(B, H // window_size, window_size, W // window_size, window_size, C)
|
| 41 |
+
windows = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(-1, window_size, window_size, C)
|
| 42 |
+
return windows
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
def window_reverse(windows, window_size, H, W):
|
| 46 |
+
"""
|
| 47 |
+
Args:
|
| 48 |
+
windows: (num_windows*B, window_size, window_size, C)
|
| 49 |
+
window_size (int): Window size
|
| 50 |
+
H (int): Height of image
|
| 51 |
+
W (int): Width of image
|
| 52 |
+
|
| 53 |
+
Returns:
|
| 54 |
+
x: (B, H, W, C)
|
| 55 |
+
"""
|
| 56 |
+
B = int(windows.shape[0] / (H * W / window_size / window_size))
|
| 57 |
+
x = windows.view(B, H // window_size, W // window_size, window_size, window_size, -1)
|
| 58 |
+
x = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(B, H, W, -1)
|
| 59 |
+
return x
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
class WindowAttention(nn.Module):
|
| 63 |
+
""" Window based multi-head self attention (W-MSA) module with relative position bias.
|
| 64 |
+
It supports both of shifted and non-shifted window.
|
| 65 |
+
|
| 66 |
+
Args:
|
| 67 |
+
dim (int): Number of input channels.
|
| 68 |
+
window_size (tuple[int]): The height and width of the window.
|
| 69 |
+
num_heads (int): Number of attention heads.
|
| 70 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 71 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set
|
| 72 |
+
attn_drop (float, optional): Dropout ratio of attention weight. Default: 0.0
|
| 73 |
+
proj_drop (float, optional): Dropout ratio of output. Default: 0.0
|
| 74 |
+
"""
|
| 75 |
+
|
| 76 |
+
def __init__(self, dim, window_size, num_heads, v_dim, qkv_bias=True, qk_scale=None, attn_drop=0., proj_drop=0.):
|
| 77 |
+
|
| 78 |
+
super().__init__()
|
| 79 |
+
self.dim = dim
|
| 80 |
+
self.window_size = window_size # Wh, Ww
|
| 81 |
+
self.num_heads = num_heads
|
| 82 |
+
head_dim = dim // num_heads
|
| 83 |
+
self.scale = qk_scale or head_dim ** -0.5
|
| 84 |
+
|
| 85 |
+
# define a parameter table of relative position bias
|
| 86 |
+
self.relative_position_bias_table = nn.Parameter(
|
| 87 |
+
torch.zeros((2 * window_size[0] - 1) * (2 * window_size[1] - 1), num_heads)) # 2*Wh-1 * 2*Ww-1, nH
|
| 88 |
+
|
| 89 |
+
# get pair-wise relative position index for each token inside the window
|
| 90 |
+
coords_h = torch.arange(self.window_size[0])
|
| 91 |
+
coords_w = torch.arange(self.window_size[1])
|
| 92 |
+
coords = torch.stack(torch.meshgrid([coords_h, coords_w])) # 2, Wh, Ww
|
| 93 |
+
coords_flatten = torch.flatten(coords, 1) # 2, Wh*Ww
|
| 94 |
+
relative_coords = coords_flatten[:, :, None] - coords_flatten[:, None, :] # 2, Wh*Ww, Wh*Ww
|
| 95 |
+
relative_coords = relative_coords.permute(1, 2, 0).contiguous() # Wh*Ww, Wh*Ww, 2
|
| 96 |
+
relative_coords[:, :, 0] += self.window_size[0] - 1 # shift to start from 0
|
| 97 |
+
relative_coords[:, :, 1] += self.window_size[1] - 1
|
| 98 |
+
relative_coords[:, :, 0] *= 2 * self.window_size[1] - 1
|
| 99 |
+
relative_position_index = relative_coords.sum(-1) # Wh*Ww, Wh*Ww
|
| 100 |
+
self.register_buffer("relative_position_index", relative_position_index)
|
| 101 |
+
|
| 102 |
+
self.qk = nn.Linear(dim, dim * 2, bias=qkv_bias)
|
| 103 |
+
self.attn_drop = nn.Dropout(attn_drop)
|
| 104 |
+
self.proj = nn.Linear(v_dim, v_dim)
|
| 105 |
+
self.proj_drop = nn.Dropout(proj_drop)
|
| 106 |
+
|
| 107 |
+
trunc_normal_(self.relative_position_bias_table, std=.02)
|
| 108 |
+
self.softmax = nn.Softmax(dim=-1)
|
| 109 |
+
|
| 110 |
+
def forward(self, x, v, mask=None):
|
| 111 |
+
""" Forward function.
|
| 112 |
+
|
| 113 |
+
Args:
|
| 114 |
+
x: input features with shape of (num_windows*B, N, C)
|
| 115 |
+
mask: (0/-inf) mask with shape of (num_windows, Wh*Ww, Wh*Ww) or None
|
| 116 |
+
"""
|
| 117 |
+
B_, N, C = x.shape
|
| 118 |
+
qk = self.qk(x).reshape(B_, N, 2, self.num_heads, C // self.num_heads).permute(2, 0, 3, 1, 4)
|
| 119 |
+
q, k = qk[0], qk[1] # make torchscript happy (cannot use tensor as tuple)
|
| 120 |
+
|
| 121 |
+
q = q * self.scale
|
| 122 |
+
attn = (q @ k.transpose(-2, -1))
|
| 123 |
+
|
| 124 |
+
relative_position_bias = self.relative_position_bias_table[self.relative_position_index.view(-1)].view(
|
| 125 |
+
self.window_size[0] * self.window_size[1], self.window_size[0] * self.window_size[1], -1) # Wh*Ww,Wh*Ww,nH
|
| 126 |
+
relative_position_bias = relative_position_bias.permute(2, 0, 1).contiguous() # nH, Wh*Ww, Wh*Ww
|
| 127 |
+
attn = attn + relative_position_bias.unsqueeze(0)
|
| 128 |
+
|
| 129 |
+
if mask is not None:
|
| 130 |
+
nW = mask.shape[0]
|
| 131 |
+
attn = attn.view(B_ // nW, nW, self.num_heads, N, N) + mask.unsqueeze(1).unsqueeze(0)
|
| 132 |
+
attn = attn.view(-1, self.num_heads, N, N)
|
| 133 |
+
attn = self.softmax(attn)
|
| 134 |
+
else:
|
| 135 |
+
attn = self.softmax(attn)
|
| 136 |
+
|
| 137 |
+
attn = self.attn_drop(attn)
|
| 138 |
+
|
| 139 |
+
# assert self.dim % v.shape[-1] == 0, "self.dim % v.shape[-1] != 0"
|
| 140 |
+
# repeat_num = self.dim // v.shape[-1]
|
| 141 |
+
# v = v.view(B_, N, self.num_heads // repeat_num, -1).transpose(1, 2).repeat(1, repeat_num, 1, 1)
|
| 142 |
+
|
| 143 |
+
assert self.dim == v.shape[-1], "self.dim != v.shape[-1]"
|
| 144 |
+
v = v.view(B_, N, self.num_heads, -1).transpose(1, 2)
|
| 145 |
+
|
| 146 |
+
x = (attn @ v).transpose(1, 2).reshape(B_, N, C)
|
| 147 |
+
x = self.proj(x)
|
| 148 |
+
x = self.proj_drop(x)
|
| 149 |
+
return x
|
| 150 |
+
|
| 151 |
+
|
| 152 |
+
class CRFBlock(nn.Module):
|
| 153 |
+
""" CRF Block.
|
| 154 |
+
|
| 155 |
+
Args:
|
| 156 |
+
dim (int): Number of input channels.
|
| 157 |
+
num_heads (int): Number of attention heads.
|
| 158 |
+
window_size (int): Window size.
|
| 159 |
+
shift_size (int): Shift size for SW-MSA.
|
| 160 |
+
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
|
| 161 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 162 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
|
| 163 |
+
drop (float, optional): Dropout rate. Default: 0.0
|
| 164 |
+
attn_drop (float, optional): Attention dropout rate. Default: 0.0
|
| 165 |
+
drop_path (float, optional): Stochastic depth rate. Default: 0.0
|
| 166 |
+
act_layer (nn.Module, optional): Activation layer. Default: nn.GELU
|
| 167 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
|
| 168 |
+
"""
|
| 169 |
+
|
| 170 |
+
def __init__(self, dim, num_heads, v_dim, window_size=7, shift_size=0,
|
| 171 |
+
mlp_ratio=4., qkv_bias=True, qk_scale=None, drop=0., attn_drop=0., drop_path=0.,
|
| 172 |
+
act_layer=nn.GELU, norm_layer=nn.LayerNorm):
|
| 173 |
+
super().__init__()
|
| 174 |
+
self.dim = dim
|
| 175 |
+
self.num_heads = num_heads
|
| 176 |
+
self.v_dim = v_dim
|
| 177 |
+
self.window_size = window_size
|
| 178 |
+
self.shift_size = shift_size
|
| 179 |
+
self.mlp_ratio = mlp_ratio
|
| 180 |
+
assert 0 <= self.shift_size < self.window_size, "shift_size must in 0-window_size"
|
| 181 |
+
|
| 182 |
+
self.norm1 = norm_layer(dim)
|
| 183 |
+
self.attn = WindowAttention(
|
| 184 |
+
dim, window_size=to_2tuple(self.window_size), num_heads=num_heads, v_dim=v_dim,
|
| 185 |
+
qkv_bias=qkv_bias, qk_scale=qk_scale, attn_drop=attn_drop, proj_drop=drop)
|
| 186 |
+
|
| 187 |
+
self.drop_path = DropPath(drop_path) if drop_path > 0. else nn.Identity()
|
| 188 |
+
self.norm2 = norm_layer(v_dim)
|
| 189 |
+
mlp_hidden_dim = int(v_dim * mlp_ratio)
|
| 190 |
+
self.mlp = Mlp(in_features=v_dim, hidden_features=mlp_hidden_dim, act_layer=act_layer, drop=drop)
|
| 191 |
+
|
| 192 |
+
self.H = None
|
| 193 |
+
self.W = None
|
| 194 |
+
|
| 195 |
+
def forward(self, x, v, mask_matrix):
|
| 196 |
+
""" Forward function.
|
| 197 |
+
|
| 198 |
+
Args:
|
| 199 |
+
x: Input feature, tensor size (B, H*W, C).
|
| 200 |
+
H, W: Spatial resolution of the input feature.
|
| 201 |
+
mask_matrix: Attention mask for cyclic shift.
|
| 202 |
+
"""
|
| 203 |
+
B, L, C = x.shape
|
| 204 |
+
H, W = self.H, self.W
|
| 205 |
+
assert L == H * W, "input feature has wrong size"
|
| 206 |
+
|
| 207 |
+
shortcut = x
|
| 208 |
+
x = self.norm1(x)
|
| 209 |
+
x = x.view(B, H, W, C)
|
| 210 |
+
|
| 211 |
+
# pad feature maps to multiples of window size
|
| 212 |
+
pad_l = pad_t = 0
|
| 213 |
+
pad_r = (self.window_size - W % self.window_size) % self.window_size
|
| 214 |
+
pad_b = (self.window_size - H % self.window_size) % self.window_size
|
| 215 |
+
x = F.pad(x, (0, 0, pad_l, pad_r, pad_t, pad_b))
|
| 216 |
+
v = F.pad(v, (0, 0, pad_l, pad_r, pad_t, pad_b))
|
| 217 |
+
_, Hp, Wp, _ = x.shape
|
| 218 |
+
|
| 219 |
+
# cyclic shift
|
| 220 |
+
if self.shift_size > 0:
|
| 221 |
+
shifted_x = torch.roll(x, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
|
| 222 |
+
shifted_v = torch.roll(v, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
|
| 223 |
+
attn_mask = mask_matrix
|
| 224 |
+
else:
|
| 225 |
+
shifted_x = x
|
| 226 |
+
shifted_v = v
|
| 227 |
+
attn_mask = None
|
| 228 |
+
|
| 229 |
+
# partition windows
|
| 230 |
+
x_windows = window_partition(shifted_x, self.window_size) # nW*B, window_size, window_size, C
|
| 231 |
+
x_windows = x_windows.view(-1, self.window_size * self.window_size, C) # nW*B, window_size*window_size, C
|
| 232 |
+
v_windows = window_partition(shifted_v, self.window_size) # nW*B, window_size, window_size, C
|
| 233 |
+
v_windows = v_windows.view(-1, self.window_size * self.window_size, v_windows.shape[-1]) # nW*B, window_size*window_size, C
|
| 234 |
+
|
| 235 |
+
# W-MSA/SW-MSA
|
| 236 |
+
attn_windows = self.attn(x_windows, v_windows, mask=attn_mask) # nW*B, window_size*window_size, C
|
| 237 |
+
|
| 238 |
+
# merge windows
|
| 239 |
+
attn_windows = attn_windows.view(-1, self.window_size, self.window_size, self.v_dim)
|
| 240 |
+
shifted_x = window_reverse(attn_windows, self.window_size, Hp, Wp) # B H' W' C
|
| 241 |
+
|
| 242 |
+
# reverse cyclic shift
|
| 243 |
+
if self.shift_size > 0:
|
| 244 |
+
x = torch.roll(shifted_x, shifts=(self.shift_size, self.shift_size), dims=(1, 2))
|
| 245 |
+
else:
|
| 246 |
+
x = shifted_x
|
| 247 |
+
|
| 248 |
+
if pad_r > 0 or pad_b > 0:
|
| 249 |
+
x = x[:, :H, :W, :].contiguous()
|
| 250 |
+
|
| 251 |
+
x = x.view(B, H * W, self.v_dim)
|
| 252 |
+
|
| 253 |
+
# FFN
|
| 254 |
+
x = shortcut + self.drop_path(x)
|
| 255 |
+
x = x + self.drop_path(self.mlp(self.norm2(x)))
|
| 256 |
+
|
| 257 |
+
return x
|
| 258 |
+
|
| 259 |
+
|
| 260 |
+
class BasicCRFLayer(nn.Module):
|
| 261 |
+
""" A basic NeWCRFs layer for one stage.
|
| 262 |
+
|
| 263 |
+
Args:
|
| 264 |
+
dim (int): Number of feature channels
|
| 265 |
+
depth (int): Depths of this stage.
|
| 266 |
+
num_heads (int): Number of attention head.
|
| 267 |
+
window_size (int): Local window size. Default: 7.
|
| 268 |
+
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
|
| 269 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 270 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
|
| 271 |
+
drop (float, optional): Dropout rate. Default: 0.0
|
| 272 |
+
attn_drop (float, optional): Attention dropout rate. Default: 0.0
|
| 273 |
+
drop_path (float | tuple[float], optional): Stochastic depth rate. Default: 0.0
|
| 274 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
|
| 275 |
+
downsample (nn.Module | None, optional): Downsample layer at the end of the layer. Default: None
|
| 276 |
+
use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
|
| 277 |
+
"""
|
| 278 |
+
|
| 279 |
+
def __init__(self,
|
| 280 |
+
dim,
|
| 281 |
+
depth,
|
| 282 |
+
num_heads,
|
| 283 |
+
v_dim,
|
| 284 |
+
window_size=7,
|
| 285 |
+
mlp_ratio=4.,
|
| 286 |
+
qkv_bias=True,
|
| 287 |
+
qk_scale=None,
|
| 288 |
+
drop=0.,
|
| 289 |
+
attn_drop=0.,
|
| 290 |
+
drop_path=0.,
|
| 291 |
+
norm_layer=nn.LayerNorm,
|
| 292 |
+
downsample=None,
|
| 293 |
+
use_checkpoint=False):
|
| 294 |
+
super().__init__()
|
| 295 |
+
self.window_size = window_size
|
| 296 |
+
self.shift_size = window_size // 2
|
| 297 |
+
self.depth = depth
|
| 298 |
+
self.use_checkpoint = use_checkpoint
|
| 299 |
+
|
| 300 |
+
# build blocks
|
| 301 |
+
self.blocks = nn.ModuleList([
|
| 302 |
+
CRFBlock(
|
| 303 |
+
dim=dim,
|
| 304 |
+
num_heads=num_heads,
|
| 305 |
+
v_dim=v_dim,
|
| 306 |
+
window_size=window_size,
|
| 307 |
+
shift_size=0 if (i % 2 == 0) else window_size // 2,
|
| 308 |
+
mlp_ratio=mlp_ratio,
|
| 309 |
+
qkv_bias=qkv_bias,
|
| 310 |
+
qk_scale=qk_scale,
|
| 311 |
+
drop=drop,
|
| 312 |
+
attn_drop=attn_drop,
|
| 313 |
+
drop_path=drop_path[i] if isinstance(drop_path, list) else drop_path,
|
| 314 |
+
norm_layer=norm_layer)
|
| 315 |
+
for i in range(depth)])
|
| 316 |
+
|
| 317 |
+
# patch merging layer
|
| 318 |
+
if downsample is not None:
|
| 319 |
+
self.downsample = downsample(dim=dim, norm_layer=norm_layer)
|
| 320 |
+
else:
|
| 321 |
+
self.downsample = None
|
| 322 |
+
|
| 323 |
+
def forward(self, x, v, H, W):
|
| 324 |
+
""" Forward function.
|
| 325 |
+
|
| 326 |
+
Args:
|
| 327 |
+
x: Input feature, tensor size (B, H*W, C).
|
| 328 |
+
H, W: Spatial resolution of the input feature.
|
| 329 |
+
"""
|
| 330 |
+
|
| 331 |
+
# calculate attention mask for SW-MSA
|
| 332 |
+
Hp = int(np.ceil(H / self.window_size)) * self.window_size
|
| 333 |
+
Wp = int(np.ceil(W / self.window_size)) * self.window_size
|
| 334 |
+
img_mask = torch.zeros((1, Hp, Wp, 1), device=x.device) # 1 Hp Wp 1
|
| 335 |
+
h_slices = (slice(0, -self.window_size),
|
| 336 |
+
slice(-self.window_size, -self.shift_size),
|
| 337 |
+
slice(-self.shift_size, None))
|
| 338 |
+
w_slices = (slice(0, -self.window_size),
|
| 339 |
+
slice(-self.window_size, -self.shift_size),
|
| 340 |
+
slice(-self.shift_size, None))
|
| 341 |
+
cnt = 0
|
| 342 |
+
for h in h_slices:
|
| 343 |
+
for w in w_slices:
|
| 344 |
+
img_mask[:, h, w, :] = cnt
|
| 345 |
+
cnt += 1
|
| 346 |
+
|
| 347 |
+
mask_windows = window_partition(img_mask, self.window_size) # nW, window_size, window_size, 1
|
| 348 |
+
mask_windows = mask_windows.view(-1, self.window_size * self.window_size)
|
| 349 |
+
attn_mask = mask_windows.unsqueeze(1) - mask_windows.unsqueeze(2)
|
| 350 |
+
attn_mask = attn_mask.masked_fill(attn_mask != 0, float(-100.0)).masked_fill(attn_mask == 0, float(0.0))
|
| 351 |
+
|
| 352 |
+
for blk in self.blocks:
|
| 353 |
+
blk.H, blk.W = H, W
|
| 354 |
+
if self.use_checkpoint:
|
| 355 |
+
x = checkpoint.checkpoint(blk, x, attn_mask)
|
| 356 |
+
else:
|
| 357 |
+
x = blk(x, v, attn_mask)
|
| 358 |
+
if self.downsample is not None:
|
| 359 |
+
x_down = self.downsample(x, H, W)
|
| 360 |
+
Wh, Ww = (H + 1) // 2, (W + 1) // 2
|
| 361 |
+
return x, H, W, x_down, Wh, Ww
|
| 362 |
+
else:
|
| 363 |
+
return x, H, W, x, H, W
|
| 364 |
+
|
| 365 |
+
|
| 366 |
+
class NewCRF(nn.Module):
|
| 367 |
+
def __init__(self,
|
| 368 |
+
input_dim=96,
|
| 369 |
+
embed_dim=96,
|
| 370 |
+
v_dim=64,
|
| 371 |
+
window_size=7,
|
| 372 |
+
num_heads=4,
|
| 373 |
+
depth=2,
|
| 374 |
+
patch_size=4,
|
| 375 |
+
in_chans=3,
|
| 376 |
+
norm_layer=nn.LayerNorm,
|
| 377 |
+
patch_norm=True):
|
| 378 |
+
super().__init__()
|
| 379 |
+
|
| 380 |
+
self.embed_dim = embed_dim
|
| 381 |
+
self.patch_norm = patch_norm
|
| 382 |
+
|
| 383 |
+
if input_dim != embed_dim:
|
| 384 |
+
self.proj_x = nn.Conv2d(input_dim, embed_dim, 3, padding=1)
|
| 385 |
+
else:
|
| 386 |
+
self.proj_x = None
|
| 387 |
+
|
| 388 |
+
if v_dim != embed_dim:
|
| 389 |
+
self.proj_v = nn.Conv2d(v_dim, embed_dim, 3, padding=1)
|
| 390 |
+
elif embed_dim % v_dim == 0:
|
| 391 |
+
self.proj_v = None
|
| 392 |
+
|
| 393 |
+
# For now, v_dim need to be equal to embed_dim, because the output of window-attn is the input of shift-window-attn
|
| 394 |
+
v_dim = embed_dim
|
| 395 |
+
assert v_dim == embed_dim
|
| 396 |
+
|
| 397 |
+
self.crf_layer = BasicCRFLayer(
|
| 398 |
+
dim=embed_dim,
|
| 399 |
+
depth=depth,
|
| 400 |
+
num_heads=num_heads,
|
| 401 |
+
v_dim=v_dim,
|
| 402 |
+
window_size=window_size,
|
| 403 |
+
mlp_ratio=4.,
|
| 404 |
+
qkv_bias=True,
|
| 405 |
+
qk_scale=None,
|
| 406 |
+
drop=0.,
|
| 407 |
+
attn_drop=0.,
|
| 408 |
+
drop_path=0.,
|
| 409 |
+
norm_layer=norm_layer,
|
| 410 |
+
downsample=None,
|
| 411 |
+
use_checkpoint=False)
|
| 412 |
+
|
| 413 |
+
layer = norm_layer(embed_dim)
|
| 414 |
+
layer_name = 'norm_crf'
|
| 415 |
+
self.add_module(layer_name, layer)
|
| 416 |
+
|
| 417 |
+
|
| 418 |
+
def forward(self, x, v):
|
| 419 |
+
if self.proj_x is not None:
|
| 420 |
+
x = self.proj_x(x)
|
| 421 |
+
if self.proj_v is not None:
|
| 422 |
+
v = self.proj_v(v)
|
| 423 |
+
|
| 424 |
+
Wh, Ww = x.size(2), x.size(3)
|
| 425 |
+
x = x.flatten(2).transpose(1, 2)
|
| 426 |
+
v = v.transpose(1, 2).transpose(2, 3)
|
| 427 |
+
|
| 428 |
+
x_out, H, W, x, Wh, Ww = self.crf_layer(x, v, Wh, Ww)
|
| 429 |
+
norm_layer = getattr(self, f'norm_crf')
|
| 430 |
+
x_out = norm_layer(x_out)
|
| 431 |
+
out = x_out.view(-1, H, W, self.embed_dim).permute(0, 3, 1, 2).contiguous()
|
| 432 |
+
|
| 433 |
+
return out
|
external/WildCamera/WildCamera/newcrfs/newcrf_utils.py
ADDED
|
@@ -0,0 +1,264 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import warnings
|
| 2 |
+
import os
|
| 3 |
+
import os.path as osp
|
| 4 |
+
import pkgutil
|
| 5 |
+
import warnings
|
| 6 |
+
from collections import OrderedDict
|
| 7 |
+
from importlib import import_module
|
| 8 |
+
|
| 9 |
+
import torch
|
| 10 |
+
import torchvision
|
| 11 |
+
import torch.nn as nn
|
| 12 |
+
from torch.utils import model_zoo
|
| 13 |
+
from torch.nn import functional as F
|
| 14 |
+
from torch.nn.parallel import DataParallel, DistributedDataParallel
|
| 15 |
+
from torch import distributed as dist
|
| 16 |
+
|
| 17 |
+
TORCH_VERSION = torch.__version__
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
def resize(input,
|
| 21 |
+
size=None,
|
| 22 |
+
scale_factor=None,
|
| 23 |
+
mode='nearest',
|
| 24 |
+
align_corners=None,
|
| 25 |
+
warning=True):
|
| 26 |
+
if warning:
|
| 27 |
+
if size is not None and align_corners:
|
| 28 |
+
input_h, input_w = tuple(int(x) for x in input.shape[2:])
|
| 29 |
+
output_h, output_w = tuple(int(x) for x in size)
|
| 30 |
+
if output_h > input_h or output_w > output_h:
|
| 31 |
+
if ((output_h > 1 and output_w > 1 and input_h > 1
|
| 32 |
+
and input_w > 1) and (output_h - 1) % (input_h - 1)
|
| 33 |
+
and (output_w - 1) % (input_w - 1)):
|
| 34 |
+
warnings.warn(
|
| 35 |
+
f'When align_corners={align_corners}, '
|
| 36 |
+
'the output would more aligned if '
|
| 37 |
+
f'input size {(input_h, input_w)} is `x+1` and '
|
| 38 |
+
f'out size {(output_h, output_w)} is `nx+1`')
|
| 39 |
+
if isinstance(size, torch.Size):
|
| 40 |
+
size = tuple(int(x) for x in size)
|
| 41 |
+
return F.interpolate(input, size, scale_factor, mode, align_corners)
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
def normal_init(module, mean=0, std=1, bias=0):
|
| 45 |
+
if hasattr(module, 'weight') and module.weight is not None:
|
| 46 |
+
nn.init.normal_(module.weight, mean, std)
|
| 47 |
+
if hasattr(module, 'bias') and module.bias is not None:
|
| 48 |
+
nn.init.constant_(module.bias, bias)
|
| 49 |
+
|
| 50 |
+
|
| 51 |
+
def is_module_wrapper(module):
|
| 52 |
+
module_wrappers = (DataParallel, DistributedDataParallel)
|
| 53 |
+
return isinstance(module, module_wrappers)
|
| 54 |
+
|
| 55 |
+
|
| 56 |
+
def get_dist_info():
|
| 57 |
+
if TORCH_VERSION < '1.0':
|
| 58 |
+
initialized = dist._initialized
|
| 59 |
+
else:
|
| 60 |
+
if dist.is_available():
|
| 61 |
+
initialized = dist.is_initialized()
|
| 62 |
+
else:
|
| 63 |
+
initialized = False
|
| 64 |
+
if initialized:
|
| 65 |
+
rank = dist.get_rank()
|
| 66 |
+
world_size = dist.get_world_size()
|
| 67 |
+
else:
|
| 68 |
+
rank = 0
|
| 69 |
+
world_size = 1
|
| 70 |
+
return rank, world_size
|
| 71 |
+
|
| 72 |
+
|
| 73 |
+
def load_state_dict(module, state_dict, strict=False, logger=None):
|
| 74 |
+
"""Load state_dict to a module.
|
| 75 |
+
|
| 76 |
+
This method is modified from :meth:`torch.nn.Module.load_state_dict`.
|
| 77 |
+
Default value for ``strict`` is set to ``False`` and the message for
|
| 78 |
+
param mismatch will be shown even if strict is False.
|
| 79 |
+
|
| 80 |
+
Args:
|
| 81 |
+
module (Module): Module that receives the state_dict.
|
| 82 |
+
state_dict (OrderedDict): Weights.
|
| 83 |
+
strict (bool): whether to strictly enforce that the keys
|
| 84 |
+
in :attr:`state_dict` match the keys returned by this module's
|
| 85 |
+
:meth:`~torch.nn.Module.state_dict` function. Default: ``False``.
|
| 86 |
+
logger (:obj:`logging.Logger`, optional): Logger to log the error
|
| 87 |
+
message. If not specified, print function will be used.
|
| 88 |
+
"""
|
| 89 |
+
unexpected_keys = []
|
| 90 |
+
all_missing_keys = []
|
| 91 |
+
err_msg = []
|
| 92 |
+
|
| 93 |
+
metadata = getattr(state_dict, '_metadata', None)
|
| 94 |
+
state_dict = state_dict.copy()
|
| 95 |
+
if metadata is not None:
|
| 96 |
+
state_dict._metadata = metadata
|
| 97 |
+
|
| 98 |
+
# use _load_from_state_dict to enable checkpoint version control
|
| 99 |
+
def load(module, prefix=''):
|
| 100 |
+
# recursively check parallel module in case that the model has a
|
| 101 |
+
# complicated structure, e.g., nn.Module(nn.Module(DDP))
|
| 102 |
+
if is_module_wrapper(module):
|
| 103 |
+
module = module.module
|
| 104 |
+
local_metadata = {} if metadata is None else metadata.get(
|
| 105 |
+
prefix[:-1], {})
|
| 106 |
+
module._load_from_state_dict(state_dict, prefix, local_metadata, True,
|
| 107 |
+
all_missing_keys, unexpected_keys,
|
| 108 |
+
err_msg)
|
| 109 |
+
for name, child in module._modules.items():
|
| 110 |
+
if child is not None:
|
| 111 |
+
load(child, prefix + name + '.')
|
| 112 |
+
|
| 113 |
+
load(module)
|
| 114 |
+
load = None # break load->load reference cycle
|
| 115 |
+
|
| 116 |
+
# ignore "num_batches_tracked" of BN layers
|
| 117 |
+
missing_keys = [
|
| 118 |
+
key for key in all_missing_keys if 'num_batches_tracked' not in key
|
| 119 |
+
]
|
| 120 |
+
|
| 121 |
+
if unexpected_keys:
|
| 122 |
+
err_msg.append('unexpected key in source '
|
| 123 |
+
f'state_dict: {", ".join(unexpected_keys)}\n')
|
| 124 |
+
if missing_keys:
|
| 125 |
+
err_msg.append(
|
| 126 |
+
f'missing keys in source state_dict: {", ".join(missing_keys)}\n')
|
| 127 |
+
|
| 128 |
+
rank, _ = get_dist_info()
|
| 129 |
+
if len(err_msg) > 0 and rank == 0:
|
| 130 |
+
err_msg.insert(
|
| 131 |
+
0, 'The model and loaded state dict do not match exactly\n')
|
| 132 |
+
err_msg = '\n'.join(err_msg)
|
| 133 |
+
if strict:
|
| 134 |
+
raise RuntimeError(err_msg)
|
| 135 |
+
elif logger is not None:
|
| 136 |
+
logger.warning(err_msg)
|
| 137 |
+
else:
|
| 138 |
+
print(err_msg)
|
| 139 |
+
|
| 140 |
+
|
| 141 |
+
def load_url_dist(url, model_dir=None):
|
| 142 |
+
"""In distributed setting, this function only download checkpoint at local
|
| 143 |
+
rank 0."""
|
| 144 |
+
rank, world_size = get_dist_info()
|
| 145 |
+
rank = int(os.environ.get('LOCAL_RANK', rank))
|
| 146 |
+
if rank == 0:
|
| 147 |
+
checkpoint = model_zoo.load_url(url, model_dir=model_dir)
|
| 148 |
+
if world_size > 1:
|
| 149 |
+
torch.distributed.barrier()
|
| 150 |
+
if rank > 0:
|
| 151 |
+
checkpoint = model_zoo.load_url(url, model_dir=model_dir)
|
| 152 |
+
return checkpoint
|
| 153 |
+
|
| 154 |
+
|
| 155 |
+
def get_torchvision_models():
|
| 156 |
+
model_urls = dict()
|
| 157 |
+
for _, name, ispkg in pkgutil.walk_packages(torchvision.models.__path__):
|
| 158 |
+
if ispkg:
|
| 159 |
+
continue
|
| 160 |
+
_zoo = import_module(f'torchvision.models.{name}')
|
| 161 |
+
if hasattr(_zoo, 'model_urls'):
|
| 162 |
+
_urls = getattr(_zoo, 'model_urls')
|
| 163 |
+
model_urls.update(_urls)
|
| 164 |
+
return model_urls
|
| 165 |
+
|
| 166 |
+
|
| 167 |
+
def _load_checkpoint(filename, map_location=None):
|
| 168 |
+
"""Load checkpoint from somewhere (modelzoo, file, url).
|
| 169 |
+
|
| 170 |
+
Args:
|
| 171 |
+
filename (str): Accept local filepath, URL, ``torchvision://xxx``,
|
| 172 |
+
``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for
|
| 173 |
+
details.
|
| 174 |
+
map_location (str | None): Same as :func:`torch.load`. Default: None.
|
| 175 |
+
|
| 176 |
+
Returns:
|
| 177 |
+
dict | OrderedDict: The loaded checkpoint. It can be either an
|
| 178 |
+
OrderedDict storing model weights or a dict containing other
|
| 179 |
+
information, which depends on the checkpoint.
|
| 180 |
+
"""
|
| 181 |
+
if filename.startswith('modelzoo://'):
|
| 182 |
+
warnings.warn('The URL scheme of "modelzoo://" is deprecated, please '
|
| 183 |
+
'use "torchvision://" instead')
|
| 184 |
+
model_urls = get_torchvision_models()
|
| 185 |
+
model_name = filename[11:]
|
| 186 |
+
checkpoint = load_url_dist(model_urls[model_name])
|
| 187 |
+
else:
|
| 188 |
+
if not osp.isfile(filename):
|
| 189 |
+
raise IOError(f'{filename} is not a checkpoint file')
|
| 190 |
+
checkpoint = torch.load(filename, map_location=map_location)
|
| 191 |
+
return checkpoint
|
| 192 |
+
|
| 193 |
+
|
| 194 |
+
def load_checkpoint(model,
|
| 195 |
+
filename,
|
| 196 |
+
map_location='cpu',
|
| 197 |
+
strict=False,
|
| 198 |
+
logger=None):
|
| 199 |
+
"""Load checkpoint from a file or URI.
|
| 200 |
+
|
| 201 |
+
Args:
|
| 202 |
+
model (Module): Module to load checkpoint.
|
| 203 |
+
filename (str): Accept local filepath, URL, ``torchvision://xxx``,
|
| 204 |
+
``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for
|
| 205 |
+
details.
|
| 206 |
+
map_location (str): Same as :func:`torch.load`.
|
| 207 |
+
strict (bool): Whether to allow different params for the model and
|
| 208 |
+
checkpoint.
|
| 209 |
+
logger (:mod:`logging.Logger` or None): The logger for error message.
|
| 210 |
+
|
| 211 |
+
Returns:
|
| 212 |
+
dict or OrderedDict: The loaded checkpoint.
|
| 213 |
+
"""
|
| 214 |
+
checkpoint = _load_checkpoint(filename, map_location)
|
| 215 |
+
# OrderedDict is a subclass of dict
|
| 216 |
+
if not isinstance(checkpoint, dict):
|
| 217 |
+
raise RuntimeError(
|
| 218 |
+
f'No state_dict found in checkpoint file {filename}')
|
| 219 |
+
# get state_dict from checkpoint
|
| 220 |
+
if 'state_dict' in checkpoint:
|
| 221 |
+
state_dict = checkpoint['state_dict']
|
| 222 |
+
elif 'model' in checkpoint:
|
| 223 |
+
state_dict = checkpoint['model']
|
| 224 |
+
else:
|
| 225 |
+
state_dict = checkpoint
|
| 226 |
+
# strip prefix of state_dict
|
| 227 |
+
if list(state_dict.keys())[0].startswith('module.'):
|
| 228 |
+
state_dict = {k[7:]: v for k, v in state_dict.items()}
|
| 229 |
+
|
| 230 |
+
# for MoBY, load model of online branch
|
| 231 |
+
if sorted(list(state_dict.keys()))[0].startswith('encoder'):
|
| 232 |
+
state_dict = {k.replace('encoder.', ''): v for k, v in state_dict.items() if k.startswith('encoder.')}
|
| 233 |
+
|
| 234 |
+
# reshape absolute position embedding
|
| 235 |
+
if state_dict.get('absolute_pos_embed') is not None:
|
| 236 |
+
absolute_pos_embed = state_dict['absolute_pos_embed']
|
| 237 |
+
N1, L, C1 = absolute_pos_embed.size()
|
| 238 |
+
N2, C2, H, W = model.absolute_pos_embed.size()
|
| 239 |
+
if N1 != N2 or C1 != C2 or L != H*W:
|
| 240 |
+
logger.warning("Error in loading absolute_pos_embed, pass")
|
| 241 |
+
else:
|
| 242 |
+
state_dict['absolute_pos_embed'] = absolute_pos_embed.view(N2, H, W, C2).permute(0, 3, 1, 2)
|
| 243 |
+
|
| 244 |
+
# interpolate position bias table if needed
|
| 245 |
+
relative_position_bias_table_keys = [k for k in state_dict.keys() if "relative_position_bias_table" in k]
|
| 246 |
+
for table_key in relative_position_bias_table_keys:
|
| 247 |
+
table_pretrained = state_dict[table_key]
|
| 248 |
+
table_current = model.state_dict()[table_key]
|
| 249 |
+
L1, nH1 = table_pretrained.size()
|
| 250 |
+
L2, nH2 = table_current.size()
|
| 251 |
+
if nH1 != nH2:
|
| 252 |
+
logger.warning(f"Error in loading {table_key}, pass")
|
| 253 |
+
else:
|
| 254 |
+
if L1 != L2:
|
| 255 |
+
S1 = int(L1 ** 0.5)
|
| 256 |
+
S2 = int(L2 ** 0.5)
|
| 257 |
+
table_pretrained_resized = F.interpolate(
|
| 258 |
+
table_pretrained.permute(1, 0).view(1, nH1, S1, S1),
|
| 259 |
+
size=(S2, S2), mode='bicubic')
|
| 260 |
+
state_dict[table_key] = table_pretrained_resized.view(nH2, L2).permute(1, 0)
|
| 261 |
+
|
| 262 |
+
# load state_dict
|
| 263 |
+
load_state_dict(model, state_dict, strict, logger)
|
| 264 |
+
return checkpoint
|
external/WildCamera/WildCamera/newcrfs/swin_transformer.py
ADDED
|
@@ -0,0 +1,619 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
import torch.nn.functional as F
|
| 4 |
+
import torch.utils.checkpoint as checkpoint
|
| 5 |
+
import numpy as np
|
| 6 |
+
from timm.models.layers import DropPath, to_2tuple, trunc_normal_
|
| 7 |
+
|
| 8 |
+
from .newcrf_utils import load_checkpoint
|
| 9 |
+
|
| 10 |
+
class Mlp(nn.Module):
|
| 11 |
+
""" Multilayer perceptron."""
|
| 12 |
+
|
| 13 |
+
def __init__(self, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, drop=0.):
|
| 14 |
+
super().__init__()
|
| 15 |
+
out_features = out_features or in_features
|
| 16 |
+
hidden_features = hidden_features or in_features
|
| 17 |
+
self.fc1 = nn.Linear(in_features, hidden_features)
|
| 18 |
+
self.act = act_layer()
|
| 19 |
+
self.fc2 = nn.Linear(hidden_features, out_features)
|
| 20 |
+
self.drop = nn.Dropout(drop)
|
| 21 |
+
|
| 22 |
+
def forward(self, x):
|
| 23 |
+
x = self.fc1(x)
|
| 24 |
+
x = self.act(x)
|
| 25 |
+
x = self.drop(x)
|
| 26 |
+
x = self.fc2(x)
|
| 27 |
+
x = self.drop(x)
|
| 28 |
+
return x
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
def window_partition(x, window_size):
|
| 32 |
+
"""
|
| 33 |
+
Args:
|
| 34 |
+
x: (B, H, W, C)
|
| 35 |
+
window_size (int): window size
|
| 36 |
+
|
| 37 |
+
Returns:
|
| 38 |
+
windows: (num_windows*B, window_size, window_size, C)
|
| 39 |
+
"""
|
| 40 |
+
B, H, W, C = x.shape
|
| 41 |
+
x = x.view(B, H // window_size, window_size, W // window_size, window_size, C)
|
| 42 |
+
windows = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(-1, window_size, window_size, C)
|
| 43 |
+
return windows
|
| 44 |
+
|
| 45 |
+
|
| 46 |
+
def window_reverse(windows, window_size, H, W):
|
| 47 |
+
"""
|
| 48 |
+
Args:
|
| 49 |
+
windows: (num_windows*B, window_size, window_size, C)
|
| 50 |
+
window_size (int): Window size
|
| 51 |
+
H (int): Height of image
|
| 52 |
+
W (int): Width of image
|
| 53 |
+
|
| 54 |
+
Returns:
|
| 55 |
+
x: (B, H, W, C)
|
| 56 |
+
"""
|
| 57 |
+
B = int(windows.shape[0] / (H * W / window_size / window_size))
|
| 58 |
+
x = windows.view(B, H // window_size, W // window_size, window_size, window_size, -1)
|
| 59 |
+
x = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(B, H, W, -1)
|
| 60 |
+
return x
|
| 61 |
+
|
| 62 |
+
|
| 63 |
+
class WindowAttention(nn.Module):
|
| 64 |
+
""" Window based multi-head self attention (W-MSA) module with relative position bias.
|
| 65 |
+
It supports both of shifted and non-shifted window.
|
| 66 |
+
|
| 67 |
+
Args:
|
| 68 |
+
dim (int): Number of input channels.
|
| 69 |
+
window_size (tuple[int]): The height and width of the window.
|
| 70 |
+
num_heads (int): Number of attention heads.
|
| 71 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 72 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set
|
| 73 |
+
attn_drop (float, optional): Dropout ratio of attention weight. Default: 0.0
|
| 74 |
+
proj_drop (float, optional): Dropout ratio of output. Default: 0.0
|
| 75 |
+
"""
|
| 76 |
+
|
| 77 |
+
def __init__(self, dim, window_size, num_heads, qkv_bias=True, qk_scale=None, attn_drop=0., proj_drop=0.):
|
| 78 |
+
|
| 79 |
+
super().__init__()
|
| 80 |
+
self.dim = dim
|
| 81 |
+
self.window_size = window_size # Wh, Ww
|
| 82 |
+
self.num_heads = num_heads
|
| 83 |
+
head_dim = dim // num_heads
|
| 84 |
+
self.scale = qk_scale or head_dim ** -0.5
|
| 85 |
+
|
| 86 |
+
# define a parameter table of relative position bias
|
| 87 |
+
self.relative_position_bias_table = nn.Parameter(
|
| 88 |
+
torch.zeros((2 * window_size[0] - 1) * (2 * window_size[1] - 1), num_heads)) # 2*Wh-1 * 2*Ww-1, nH
|
| 89 |
+
|
| 90 |
+
# get pair-wise relative position index for each token inside the window
|
| 91 |
+
coords_h = torch.arange(self.window_size[0])
|
| 92 |
+
coords_w = torch.arange(self.window_size[1])
|
| 93 |
+
coords = torch.stack(torch.meshgrid([coords_h, coords_w])) # 2, Wh, Ww
|
| 94 |
+
coords_flatten = torch.flatten(coords, 1) # 2, Wh*Ww
|
| 95 |
+
relative_coords = coords_flatten[:, :, None] - coords_flatten[:, None, :] # 2, Wh*Ww, Wh*Ww
|
| 96 |
+
relative_coords = relative_coords.permute(1, 2, 0).contiguous() # Wh*Ww, Wh*Ww, 2
|
| 97 |
+
relative_coords[:, :, 0] += self.window_size[0] - 1 # shift to start from 0
|
| 98 |
+
relative_coords[:, :, 1] += self.window_size[1] - 1
|
| 99 |
+
relative_coords[:, :, 0] *= 2 * self.window_size[1] - 1
|
| 100 |
+
relative_position_index = relative_coords.sum(-1) # Wh*Ww, Wh*Ww
|
| 101 |
+
self.register_buffer("relative_position_index", relative_position_index)
|
| 102 |
+
|
| 103 |
+
self.qkv = nn.Linear(dim, dim * 3, bias=qkv_bias)
|
| 104 |
+
self.attn_drop = nn.Dropout(attn_drop)
|
| 105 |
+
self.proj = nn.Linear(dim, dim)
|
| 106 |
+
self.proj_drop = nn.Dropout(proj_drop)
|
| 107 |
+
|
| 108 |
+
trunc_normal_(self.relative_position_bias_table, std=.02)
|
| 109 |
+
self.softmax = nn.Softmax(dim=-1)
|
| 110 |
+
|
| 111 |
+
def forward(self, x, mask=None):
|
| 112 |
+
""" Forward function.
|
| 113 |
+
|
| 114 |
+
Args:
|
| 115 |
+
x: input features with shape of (num_windows*B, N, C)
|
| 116 |
+
mask: (0/-inf) mask with shape of (num_windows, Wh*Ww, Wh*Ww) or None
|
| 117 |
+
"""
|
| 118 |
+
B_, N, C = x.shape
|
| 119 |
+
qkv = self.qkv(x).reshape(B_, N, 3, self.num_heads, C // self.num_heads).permute(2, 0, 3, 1, 4)
|
| 120 |
+
q, k, v = qkv[0], qkv[1], qkv[2] # make torchscript happy (cannot use tensor as tuple)
|
| 121 |
+
|
| 122 |
+
q = q * self.scale
|
| 123 |
+
attn = (q @ k.transpose(-2, -1))
|
| 124 |
+
|
| 125 |
+
relative_position_bias = self.relative_position_bias_table[self.relative_position_index.view(-1)].view(
|
| 126 |
+
self.window_size[0] * self.window_size[1], self.window_size[0] * self.window_size[1], -1) # Wh*Ww,Wh*Ww,nH
|
| 127 |
+
relative_position_bias = relative_position_bias.permute(2, 0, 1).contiguous() # nH, Wh*Ww, Wh*Ww
|
| 128 |
+
attn = attn + relative_position_bias.unsqueeze(0)
|
| 129 |
+
|
| 130 |
+
if mask is not None:
|
| 131 |
+
nW = mask.shape[0]
|
| 132 |
+
attn = attn.view(B_ // nW, nW, self.num_heads, N, N) + mask.unsqueeze(1).unsqueeze(0)
|
| 133 |
+
attn = attn.view(-1, self.num_heads, N, N)
|
| 134 |
+
attn = self.softmax(attn)
|
| 135 |
+
else:
|
| 136 |
+
attn = self.softmax(attn)
|
| 137 |
+
|
| 138 |
+
attn = self.attn_drop(attn)
|
| 139 |
+
|
| 140 |
+
x = (attn @ v).transpose(1, 2).reshape(B_, N, C)
|
| 141 |
+
x = self.proj(x)
|
| 142 |
+
x = self.proj_drop(x)
|
| 143 |
+
return x
|
| 144 |
+
|
| 145 |
+
|
| 146 |
+
class SwinTransformerBlock(nn.Module):
|
| 147 |
+
""" Swin Transformer Block.
|
| 148 |
+
|
| 149 |
+
Args:
|
| 150 |
+
dim (int): Number of input channels.
|
| 151 |
+
num_heads (int): Number of attention heads.
|
| 152 |
+
window_size (int): Window size.
|
| 153 |
+
shift_size (int): Shift size for SW-MSA.
|
| 154 |
+
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
|
| 155 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 156 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
|
| 157 |
+
drop (float, optional): Dropout rate. Default: 0.0
|
| 158 |
+
attn_drop (float, optional): Attention dropout rate. Default: 0.0
|
| 159 |
+
drop_path (float, optional): Stochastic depth rate. Default: 0.0
|
| 160 |
+
act_layer (nn.Module, optional): Activation layer. Default: nn.GELU
|
| 161 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
|
| 162 |
+
"""
|
| 163 |
+
|
| 164 |
+
def __init__(self, dim, num_heads, window_size=7, shift_size=0,
|
| 165 |
+
mlp_ratio=4., qkv_bias=True, qk_scale=None, drop=0., attn_drop=0., drop_path=0.,
|
| 166 |
+
act_layer=nn.GELU, norm_layer=nn.LayerNorm):
|
| 167 |
+
super().__init__()
|
| 168 |
+
self.dim = dim
|
| 169 |
+
self.num_heads = num_heads
|
| 170 |
+
self.window_size = window_size
|
| 171 |
+
self.shift_size = shift_size
|
| 172 |
+
self.mlp_ratio = mlp_ratio
|
| 173 |
+
assert 0 <= self.shift_size < self.window_size, "shift_size must in 0-window_size"
|
| 174 |
+
|
| 175 |
+
self.norm1 = norm_layer(dim)
|
| 176 |
+
self.attn = WindowAttention(
|
| 177 |
+
dim, window_size=to_2tuple(self.window_size), num_heads=num_heads,
|
| 178 |
+
qkv_bias=qkv_bias, qk_scale=qk_scale, attn_drop=attn_drop, proj_drop=drop)
|
| 179 |
+
|
| 180 |
+
self.drop_path = DropPath(drop_path) if drop_path > 0. else nn.Identity()
|
| 181 |
+
self.norm2 = norm_layer(dim)
|
| 182 |
+
mlp_hidden_dim = int(dim * mlp_ratio)
|
| 183 |
+
self.mlp = Mlp(in_features=dim, hidden_features=mlp_hidden_dim, act_layer=act_layer, drop=drop)
|
| 184 |
+
|
| 185 |
+
self.H = None
|
| 186 |
+
self.W = None
|
| 187 |
+
|
| 188 |
+
def forward(self, x, mask_matrix):
|
| 189 |
+
""" Forward function.
|
| 190 |
+
|
| 191 |
+
Args:
|
| 192 |
+
x: Input feature, tensor size (B, H*W, C).
|
| 193 |
+
H, W: Spatial resolution of the input feature.
|
| 194 |
+
mask_matrix: Attention mask for cyclic shift.
|
| 195 |
+
"""
|
| 196 |
+
B, L, C = x.shape
|
| 197 |
+
H, W = self.H, self.W
|
| 198 |
+
assert L == H * W, "input feature has wrong size"
|
| 199 |
+
|
| 200 |
+
shortcut = x
|
| 201 |
+
x = self.norm1(x)
|
| 202 |
+
x = x.view(B, H, W, C)
|
| 203 |
+
|
| 204 |
+
# pad feature maps to multiples of window size
|
| 205 |
+
pad_l = pad_t = 0
|
| 206 |
+
pad_r = (self.window_size - W % self.window_size) % self.window_size
|
| 207 |
+
pad_b = (self.window_size - H % self.window_size) % self.window_size
|
| 208 |
+
x = F.pad(x, (0, 0, pad_l, pad_r, pad_t, pad_b))
|
| 209 |
+
_, Hp, Wp, _ = x.shape
|
| 210 |
+
|
| 211 |
+
# cyclic shift
|
| 212 |
+
if self.shift_size > 0:
|
| 213 |
+
shifted_x = torch.roll(x, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
|
| 214 |
+
attn_mask = mask_matrix
|
| 215 |
+
else:
|
| 216 |
+
shifted_x = x
|
| 217 |
+
attn_mask = None
|
| 218 |
+
|
| 219 |
+
# partition windows
|
| 220 |
+
x_windows = window_partition(shifted_x, self.window_size) # nW*B, window_size, window_size, C
|
| 221 |
+
x_windows = x_windows.view(-1, self.window_size * self.window_size, C) # nW*B, window_size*window_size, C
|
| 222 |
+
|
| 223 |
+
# W-MSA/SW-MSA
|
| 224 |
+
attn_windows = self.attn(x_windows, mask=attn_mask) # nW*B, window_size*window_size, C
|
| 225 |
+
|
| 226 |
+
# merge windows
|
| 227 |
+
attn_windows = attn_windows.view(-1, self.window_size, self.window_size, C)
|
| 228 |
+
shifted_x = window_reverse(attn_windows, self.window_size, Hp, Wp) # B H' W' C
|
| 229 |
+
|
| 230 |
+
# reverse cyclic shift
|
| 231 |
+
if self.shift_size > 0:
|
| 232 |
+
x = torch.roll(shifted_x, shifts=(self.shift_size, self.shift_size), dims=(1, 2))
|
| 233 |
+
else:
|
| 234 |
+
x = shifted_x
|
| 235 |
+
|
| 236 |
+
if pad_r > 0 or pad_b > 0:
|
| 237 |
+
x = x[:, :H, :W, :].contiguous()
|
| 238 |
+
|
| 239 |
+
x = x.view(B, H * W, C)
|
| 240 |
+
|
| 241 |
+
# FFN
|
| 242 |
+
x = shortcut + self.drop_path(x)
|
| 243 |
+
x = x + self.drop_path(self.mlp(self.norm2(x)))
|
| 244 |
+
|
| 245 |
+
return x
|
| 246 |
+
|
| 247 |
+
|
| 248 |
+
class PatchMerging(nn.Module):
|
| 249 |
+
""" Patch Merging Layer
|
| 250 |
+
|
| 251 |
+
Args:
|
| 252 |
+
dim (int): Number of input channels.
|
| 253 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
|
| 254 |
+
"""
|
| 255 |
+
def __init__(self, dim, norm_layer=nn.LayerNorm):
|
| 256 |
+
super().__init__()
|
| 257 |
+
self.dim = dim
|
| 258 |
+
self.reduction = nn.Linear(4 * dim, 2 * dim, bias=False)
|
| 259 |
+
self.norm = norm_layer(4 * dim)
|
| 260 |
+
|
| 261 |
+
def forward(self, x, H, W):
|
| 262 |
+
""" Forward function.
|
| 263 |
+
|
| 264 |
+
Args:
|
| 265 |
+
x: Input feature, tensor size (B, H*W, C).
|
| 266 |
+
H, W: Spatial resolution of the input feature.
|
| 267 |
+
"""
|
| 268 |
+
B, L, C = x.shape
|
| 269 |
+
assert L == H * W, "input feature has wrong size"
|
| 270 |
+
|
| 271 |
+
x = x.view(B, H, W, C)
|
| 272 |
+
|
| 273 |
+
# padding
|
| 274 |
+
pad_input = (H % 2 == 1) or (W % 2 == 1)
|
| 275 |
+
if pad_input:
|
| 276 |
+
x = F.pad(x, (0, 0, 0, W % 2, 0, H % 2))
|
| 277 |
+
|
| 278 |
+
x0 = x[:, 0::2, 0::2, :] # B H/2 W/2 C
|
| 279 |
+
x1 = x[:, 1::2, 0::2, :] # B H/2 W/2 C
|
| 280 |
+
x2 = x[:, 0::2, 1::2, :] # B H/2 W/2 C
|
| 281 |
+
x3 = x[:, 1::2, 1::2, :] # B H/2 W/2 C
|
| 282 |
+
x = torch.cat([x0, x1, x2, x3], -1) # B H/2 W/2 4*C
|
| 283 |
+
x = x.view(B, -1, 4 * C) # B H/2*W/2 4*C
|
| 284 |
+
|
| 285 |
+
x = self.norm(x)
|
| 286 |
+
x = self.reduction(x)
|
| 287 |
+
|
| 288 |
+
return x
|
| 289 |
+
|
| 290 |
+
|
| 291 |
+
class BasicLayer(nn.Module):
|
| 292 |
+
""" A basic Swin Transformer layer for one stage.
|
| 293 |
+
|
| 294 |
+
Args:
|
| 295 |
+
dim (int): Number of feature channels
|
| 296 |
+
depth (int): Depths of this stage.
|
| 297 |
+
num_heads (int): Number of attention head.
|
| 298 |
+
window_size (int): Local window size. Default: 7.
|
| 299 |
+
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
|
| 300 |
+
qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
|
| 301 |
+
qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
|
| 302 |
+
drop (float, optional): Dropout rate. Default: 0.0
|
| 303 |
+
attn_drop (float, optional): Attention dropout rate. Default: 0.0
|
| 304 |
+
drop_path (float | tuple[float], optional): Stochastic depth rate. Default: 0.0
|
| 305 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
|
| 306 |
+
downsample (nn.Module | None, optional): Downsample layer at the end of the layer. Default: None
|
| 307 |
+
use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
|
| 308 |
+
"""
|
| 309 |
+
|
| 310 |
+
def __init__(self,
|
| 311 |
+
dim,
|
| 312 |
+
depth,
|
| 313 |
+
num_heads,
|
| 314 |
+
window_size=7,
|
| 315 |
+
mlp_ratio=4.,
|
| 316 |
+
qkv_bias=True,
|
| 317 |
+
qk_scale=None,
|
| 318 |
+
drop=0.,
|
| 319 |
+
attn_drop=0.,
|
| 320 |
+
drop_path=0.,
|
| 321 |
+
norm_layer=nn.LayerNorm,
|
| 322 |
+
downsample=None,
|
| 323 |
+
use_checkpoint=False):
|
| 324 |
+
super().__init__()
|
| 325 |
+
self.window_size = window_size
|
| 326 |
+
self.shift_size = window_size // 2
|
| 327 |
+
self.depth = depth
|
| 328 |
+
self.use_checkpoint = use_checkpoint
|
| 329 |
+
|
| 330 |
+
# build blocks
|
| 331 |
+
self.blocks = nn.ModuleList([
|
| 332 |
+
SwinTransformerBlock(
|
| 333 |
+
dim=dim,
|
| 334 |
+
num_heads=num_heads,
|
| 335 |
+
window_size=window_size,
|
| 336 |
+
shift_size=0 if (i % 2 == 0) else window_size // 2,
|
| 337 |
+
mlp_ratio=mlp_ratio,
|
| 338 |
+
qkv_bias=qkv_bias,
|
| 339 |
+
qk_scale=qk_scale,
|
| 340 |
+
drop=drop,
|
| 341 |
+
attn_drop=attn_drop,
|
| 342 |
+
drop_path=drop_path[i] if isinstance(drop_path, list) else drop_path,
|
| 343 |
+
norm_layer=norm_layer)
|
| 344 |
+
for i in range(depth)])
|
| 345 |
+
|
| 346 |
+
# patch merging layer
|
| 347 |
+
if downsample is not None:
|
| 348 |
+
self.downsample = downsample(dim=dim, norm_layer=norm_layer)
|
| 349 |
+
else:
|
| 350 |
+
self.downsample = None
|
| 351 |
+
|
| 352 |
+
def forward(self, x, H, W):
|
| 353 |
+
""" Forward function.
|
| 354 |
+
|
| 355 |
+
Args:
|
| 356 |
+
x: Input feature, tensor size (B, H*W, C).
|
| 357 |
+
H, W: Spatial resolution of the input feature.
|
| 358 |
+
"""
|
| 359 |
+
|
| 360 |
+
# calculate attention mask for SW-MSA
|
| 361 |
+
Hp = int(np.ceil(H / self.window_size)) * self.window_size
|
| 362 |
+
Wp = int(np.ceil(W / self.window_size)) * self.window_size
|
| 363 |
+
img_mask = torch.zeros((1, Hp, Wp, 1), device=x.device) # 1 Hp Wp 1
|
| 364 |
+
h_slices = (slice(0, -self.window_size),
|
| 365 |
+
slice(-self.window_size, -self.shift_size),
|
| 366 |
+
slice(-self.shift_size, None))
|
| 367 |
+
w_slices = (slice(0, -self.window_size),
|
| 368 |
+
slice(-self.window_size, -self.shift_size),
|
| 369 |
+
slice(-self.shift_size, None))
|
| 370 |
+
cnt = 0
|
| 371 |
+
for h in h_slices:
|
| 372 |
+
for w in w_slices:
|
| 373 |
+
img_mask[:, h, w, :] = cnt
|
| 374 |
+
cnt += 1
|
| 375 |
+
|
| 376 |
+
mask_windows = window_partition(img_mask, self.window_size) # nW, window_size, window_size, 1
|
| 377 |
+
mask_windows = mask_windows.view(-1, self.window_size * self.window_size)
|
| 378 |
+
attn_mask = mask_windows.unsqueeze(1) - mask_windows.unsqueeze(2)
|
| 379 |
+
attn_mask = attn_mask.masked_fill(attn_mask != 0, float(-100.0)).masked_fill(attn_mask == 0, float(0.0))
|
| 380 |
+
|
| 381 |
+
for blk in self.blocks:
|
| 382 |
+
blk.H, blk.W = H, W
|
| 383 |
+
if self.use_checkpoint:
|
| 384 |
+
x = checkpoint.checkpoint(blk, x, attn_mask)
|
| 385 |
+
else:
|
| 386 |
+
x = blk(x, attn_mask)
|
| 387 |
+
if self.downsample is not None:
|
| 388 |
+
x_down = self.downsample(x, H, W)
|
| 389 |
+
Wh, Ww = (H + 1) // 2, (W + 1) // 2
|
| 390 |
+
return x, H, W, x_down, Wh, Ww
|
| 391 |
+
else:
|
| 392 |
+
return x, H, W, x, H, W
|
| 393 |
+
|
| 394 |
+
|
| 395 |
+
class PatchEmbed(nn.Module):
|
| 396 |
+
""" Image to Patch Embedding
|
| 397 |
+
|
| 398 |
+
Args:
|
| 399 |
+
patch_size (int): Patch token size. Default: 4.
|
| 400 |
+
in_chans (int): Number of input image channels. Default: 3.
|
| 401 |
+
embed_dim (int): Number of linear projection output channels. Default: 96.
|
| 402 |
+
norm_layer (nn.Module, optional): Normalization layer. Default: None
|
| 403 |
+
"""
|
| 404 |
+
|
| 405 |
+
def __init__(self, patch_size=4, in_chans=3, embed_dim=96, norm_layer=None):
|
| 406 |
+
super().__init__()
|
| 407 |
+
patch_size = to_2tuple(patch_size)
|
| 408 |
+
self.patch_size = patch_size
|
| 409 |
+
|
| 410 |
+
self.in_chans = in_chans
|
| 411 |
+
self.embed_dim = embed_dim
|
| 412 |
+
|
| 413 |
+
self.proj = nn.Conv2d(in_chans, embed_dim, kernel_size=patch_size, stride=patch_size)
|
| 414 |
+
if norm_layer is not None:
|
| 415 |
+
self.norm = norm_layer(embed_dim)
|
| 416 |
+
else:
|
| 417 |
+
self.norm = None
|
| 418 |
+
|
| 419 |
+
def forward(self, x):
|
| 420 |
+
"""Forward function."""
|
| 421 |
+
# padding
|
| 422 |
+
_, _, H, W = x.size()
|
| 423 |
+
if W % self.patch_size[1] != 0:
|
| 424 |
+
x = F.pad(x, (0, self.patch_size[1] - W % self.patch_size[1]))
|
| 425 |
+
if H % self.patch_size[0] != 0:
|
| 426 |
+
x = F.pad(x, (0, 0, 0, self.patch_size[0] - H % self.patch_size[0]))
|
| 427 |
+
|
| 428 |
+
x = self.proj(x) # B C Wh Ww
|
| 429 |
+
if self.norm is not None:
|
| 430 |
+
Wh, Ww = x.size(2), x.size(3)
|
| 431 |
+
x = x.flatten(2).transpose(1, 2)
|
| 432 |
+
x = self.norm(x)
|
| 433 |
+
x = x.transpose(1, 2).view(-1, self.embed_dim, Wh, Ww)
|
| 434 |
+
|
| 435 |
+
return x
|
| 436 |
+
|
| 437 |
+
|
| 438 |
+
class SwinTransformer(nn.Module):
|
| 439 |
+
""" Swin Transformer backbone.
|
| 440 |
+
A PyTorch impl of : `Swin Transformer: Hierarchical Vision Transformer using Shifted Windows` -
|
| 441 |
+
https://arxiv.org/pdf/2103.14030
|
| 442 |
+
|
| 443 |
+
Args:
|
| 444 |
+
pretrain_img_size (int): Input image size for training the pretrained model,
|
| 445 |
+
used in absolute postion embedding. Default 224.
|
| 446 |
+
patch_size (int | tuple(int)): Patch size. Default: 4.
|
| 447 |
+
in_chans (int): Number of input image channels. Default: 3.
|
| 448 |
+
embed_dim (int): Number of linear projection output channels. Default: 96.
|
| 449 |
+
depths (tuple[int]): Depths of each Swin Transformer stage.
|
| 450 |
+
num_heads (tuple[int]): Number of attention head of each stage.
|
| 451 |
+
window_size (int): Window size. Default: 7.
|
| 452 |
+
mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
|
| 453 |
+
qkv_bias (bool): If True, add a learnable bias to query, key, value. Default: True
|
| 454 |
+
qk_scale (float): Override default qk scale of head_dim ** -0.5 if set.
|
| 455 |
+
drop_rate (float): Dropout rate.
|
| 456 |
+
attn_drop_rate (float): Attention dropout rate. Default: 0.
|
| 457 |
+
drop_path_rate (float): Stochastic depth rate. Default: 0.2.
|
| 458 |
+
norm_layer (nn.Module): Normalization layer. Default: nn.LayerNorm.
|
| 459 |
+
ape (bool): If True, add absolute position embedding to the patch embedding. Default: False.
|
| 460 |
+
patch_norm (bool): If True, add normalization after patch embedding. Default: True.
|
| 461 |
+
out_indices (Sequence[int]): Output from which stages.
|
| 462 |
+
frozen_stages (int): Stages to be frozen (stop grad and set eval mode).
|
| 463 |
+
-1 means not freezing any parameters.
|
| 464 |
+
use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
|
| 465 |
+
"""
|
| 466 |
+
|
| 467 |
+
def __init__(self,
|
| 468 |
+
pretrain_img_size=224,
|
| 469 |
+
patch_size=4,
|
| 470 |
+
in_chans=3,
|
| 471 |
+
embed_dim=96,
|
| 472 |
+
depths=[2, 2, 6, 2],
|
| 473 |
+
num_heads=[3, 6, 12, 24],
|
| 474 |
+
window_size=7,
|
| 475 |
+
mlp_ratio=4.,
|
| 476 |
+
qkv_bias=True,
|
| 477 |
+
qk_scale=None,
|
| 478 |
+
drop_rate=0.,
|
| 479 |
+
attn_drop_rate=0.,
|
| 480 |
+
drop_path_rate=0.2,
|
| 481 |
+
norm_layer=nn.LayerNorm,
|
| 482 |
+
ape=False,
|
| 483 |
+
patch_norm=True,
|
| 484 |
+
out_indices=(0, 1, 2, 3),
|
| 485 |
+
frozen_stages=-1,
|
| 486 |
+
use_checkpoint=False):
|
| 487 |
+
super().__init__()
|
| 488 |
+
|
| 489 |
+
self.pretrain_img_size = pretrain_img_size
|
| 490 |
+
self.num_layers = len(depths)
|
| 491 |
+
self.embed_dim = embed_dim
|
| 492 |
+
self.ape = ape
|
| 493 |
+
self.patch_norm = patch_norm
|
| 494 |
+
self.out_indices = out_indices
|
| 495 |
+
self.frozen_stages = frozen_stages
|
| 496 |
+
|
| 497 |
+
# split image into non-overlapping patches
|
| 498 |
+
self.patch_embed = PatchEmbed(
|
| 499 |
+
patch_size=patch_size, in_chans=in_chans, embed_dim=embed_dim,
|
| 500 |
+
norm_layer=norm_layer if self.patch_norm else None)
|
| 501 |
+
|
| 502 |
+
# absolute position embedding
|
| 503 |
+
if self.ape:
|
| 504 |
+
pretrain_img_size = to_2tuple(pretrain_img_size)
|
| 505 |
+
patch_size = to_2tuple(patch_size)
|
| 506 |
+
patches_resolution = [pretrain_img_size[0] // patch_size[0], pretrain_img_size[1] // patch_size[1]]
|
| 507 |
+
|
| 508 |
+
self.absolute_pos_embed = nn.Parameter(torch.zeros(1, embed_dim, patches_resolution[0], patches_resolution[1]))
|
| 509 |
+
trunc_normal_(self.absolute_pos_embed, std=.02)
|
| 510 |
+
|
| 511 |
+
self.pos_drop = nn.Dropout(p=drop_rate)
|
| 512 |
+
|
| 513 |
+
# stochastic depth
|
| 514 |
+
dpr = [x.item() for x in torch.linspace(0, drop_path_rate, sum(depths))] # stochastic depth decay rule
|
| 515 |
+
|
| 516 |
+
# build layers
|
| 517 |
+
self.layers = nn.ModuleList()
|
| 518 |
+
for i_layer in range(self.num_layers):
|
| 519 |
+
layer = BasicLayer(
|
| 520 |
+
dim=int(embed_dim * 2 ** i_layer),
|
| 521 |
+
depth=depths[i_layer],
|
| 522 |
+
num_heads=num_heads[i_layer],
|
| 523 |
+
window_size=window_size,
|
| 524 |
+
mlp_ratio=mlp_ratio,
|
| 525 |
+
qkv_bias=qkv_bias,
|
| 526 |
+
qk_scale=qk_scale,
|
| 527 |
+
drop=drop_rate,
|
| 528 |
+
attn_drop=attn_drop_rate,
|
| 529 |
+
drop_path=dpr[sum(depths[:i_layer]):sum(depths[:i_layer + 1])],
|
| 530 |
+
norm_layer=norm_layer,
|
| 531 |
+
downsample=PatchMerging if (i_layer < self.num_layers - 1) else None,
|
| 532 |
+
use_checkpoint=use_checkpoint)
|
| 533 |
+
self.layers.append(layer)
|
| 534 |
+
|
| 535 |
+
num_features = [int(embed_dim * 2 ** i) for i in range(self.num_layers)]
|
| 536 |
+
self.num_features = num_features
|
| 537 |
+
|
| 538 |
+
# add a norm layer for each output
|
| 539 |
+
for i_layer in out_indices:
|
| 540 |
+
layer = norm_layer(num_features[i_layer])
|
| 541 |
+
layer_name = f'norm{i_layer}'
|
| 542 |
+
self.add_module(layer_name, layer)
|
| 543 |
+
|
| 544 |
+
self._freeze_stages()
|
| 545 |
+
|
| 546 |
+
def _freeze_stages(self):
|
| 547 |
+
if self.frozen_stages >= 0:
|
| 548 |
+
self.patch_embed.eval()
|
| 549 |
+
for param in self.patch_embed.parameters():
|
| 550 |
+
param.requires_grad = False
|
| 551 |
+
|
| 552 |
+
if self.frozen_stages >= 1 and self.ape:
|
| 553 |
+
self.absolute_pos_embed.requires_grad = False
|
| 554 |
+
|
| 555 |
+
if self.frozen_stages >= 2:
|
| 556 |
+
self.pos_drop.eval()
|
| 557 |
+
for i in range(0, self.frozen_stages - 1):
|
| 558 |
+
m = self.layers[i]
|
| 559 |
+
m.eval()
|
| 560 |
+
for param in m.parameters():
|
| 561 |
+
param.requires_grad = False
|
| 562 |
+
|
| 563 |
+
def init_weights(self, pretrained=None):
|
| 564 |
+
"""Initialize the weights in backbone.
|
| 565 |
+
|
| 566 |
+
Args:
|
| 567 |
+
pretrained (str, optional): Path to pre-trained weights.
|
| 568 |
+
Defaults to None.
|
| 569 |
+
"""
|
| 570 |
+
|
| 571 |
+
def _init_weights(m):
|
| 572 |
+
if isinstance(m, nn.Linear):
|
| 573 |
+
trunc_normal_(m.weight, std=.02)
|
| 574 |
+
if isinstance(m, nn.Linear) and m.bias is not None:
|
| 575 |
+
nn.init.constant_(m.bias, 0)
|
| 576 |
+
elif isinstance(m, nn.LayerNorm):
|
| 577 |
+
nn.init.constant_(m.bias, 0)
|
| 578 |
+
nn.init.constant_(m.weight, 1.0)
|
| 579 |
+
|
| 580 |
+
if isinstance(pretrained, str):
|
| 581 |
+
self.apply(_init_weights)
|
| 582 |
+
# logger = get_root_logger()
|
| 583 |
+
load_checkpoint(self, pretrained, strict=False)
|
| 584 |
+
elif pretrained is None:
|
| 585 |
+
self.apply(_init_weights)
|
| 586 |
+
else:
|
| 587 |
+
raise TypeError('pretrained must be a str or None')
|
| 588 |
+
|
| 589 |
+
def forward(self, x):
|
| 590 |
+
"""Forward function."""
|
| 591 |
+
x = self.patch_embed(x)
|
| 592 |
+
|
| 593 |
+
Wh, Ww = x.size(2), x.size(3)
|
| 594 |
+
if self.ape:
|
| 595 |
+
# interpolate the position embedding to the corresponding size
|
| 596 |
+
absolute_pos_embed = F.interpolate(self.absolute_pos_embed, size=(Wh, Ww), mode='bicubic')
|
| 597 |
+
x = (x + absolute_pos_embed).flatten(2).transpose(1, 2) # B Wh*Ww C
|
| 598 |
+
else:
|
| 599 |
+
x = x.flatten(2).transpose(1, 2)
|
| 600 |
+
x = self.pos_drop(x)
|
| 601 |
+
|
| 602 |
+
outs = []
|
| 603 |
+
for i in range(self.num_layers):
|
| 604 |
+
layer = self.layers[i]
|
| 605 |
+
x_out, H, W, x, Wh, Ww = layer(x, Wh, Ww)
|
| 606 |
+
|
| 607 |
+
if i in self.out_indices:
|
| 608 |
+
norm_layer = getattr(self, f'norm{i}')
|
| 609 |
+
x_out = norm_layer(x_out)
|
| 610 |
+
|
| 611 |
+
out = x_out.view(-1, H, W, self.num_features[i]).permute(0, 3, 1, 2).contiguous()
|
| 612 |
+
outs.append(out)
|
| 613 |
+
|
| 614 |
+
return tuple(outs)
|
| 615 |
+
|
| 616 |
+
def train(self, mode=True):
|
| 617 |
+
"""Convert the model into training mode while keep layers freezed."""
|
| 618 |
+
super(SwinTransformer, self).train(mode)
|
| 619 |
+
self._freeze_stages()
|
external/WildCamera/WildCamera/newcrfs/uper_crf_head.py
ADDED
|
@@ -0,0 +1,350 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import torch
|
| 2 |
+
import torch.nn as nn
|
| 3 |
+
|
| 4 |
+
from mmcv.cnn import ConvModule
|
| 5 |
+
from .newcrf_utils import resize
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
class PPM(nn.ModuleList):
|
| 9 |
+
"""Pooling Pyramid Module used in PSPNet.
|
| 10 |
+
|
| 11 |
+
Args:
|
| 12 |
+
pool_scales (tuple[int]): Pooling scales used in Pooling Pyramid
|
| 13 |
+
Module.
|
| 14 |
+
in_channels (int): Input channels.
|
| 15 |
+
channels (int): Channels after modules, before conv_seg.
|
| 16 |
+
conv_cfg (dict|None): Config of conv layers.
|
| 17 |
+
norm_cfg (dict|None): Config of norm layers.
|
| 18 |
+
act_cfg (dict): Config of activation layers.
|
| 19 |
+
align_corners (bool): align_corners argument of F.interpolate.
|
| 20 |
+
"""
|
| 21 |
+
|
| 22 |
+
def __init__(self, pool_scales, in_channels, channels, conv_cfg, norm_cfg,
|
| 23 |
+
act_cfg, align_corners):
|
| 24 |
+
super(PPM, self).__init__()
|
| 25 |
+
self.pool_scales = pool_scales
|
| 26 |
+
self.align_corners = align_corners
|
| 27 |
+
self.in_channels = in_channels
|
| 28 |
+
self.channels = channels
|
| 29 |
+
self.conv_cfg = conv_cfg
|
| 30 |
+
self.norm_cfg = norm_cfg
|
| 31 |
+
self.act_cfg = act_cfg
|
| 32 |
+
for pool_scale in pool_scales:
|
| 33 |
+
# == if batch size = 1, BN is not supported, change to GN
|
| 34 |
+
if pool_scale == 1: norm_cfg = dict(type='GN', requires_grad=True, num_groups=256)
|
| 35 |
+
self.append(
|
| 36 |
+
nn.Sequential(
|
| 37 |
+
nn.AdaptiveAvgPool2d(pool_scale),
|
| 38 |
+
ConvModule(
|
| 39 |
+
self.in_channels,
|
| 40 |
+
self.channels,
|
| 41 |
+
1,
|
| 42 |
+
conv_cfg=self.conv_cfg,
|
| 43 |
+
norm_cfg=norm_cfg,
|
| 44 |
+
act_cfg=self.act_cfg)))
|
| 45 |
+
|
| 46 |
+
def forward(self, x):
|
| 47 |
+
"""Forward function."""
|
| 48 |
+
ppm_outs = []
|
| 49 |
+
for ppm in self:
|
| 50 |
+
ppm_out = ppm(x)
|
| 51 |
+
upsampled_ppm_out = resize(
|
| 52 |
+
ppm_out,
|
| 53 |
+
size=x.size()[2:],
|
| 54 |
+
mode='bilinear',
|
| 55 |
+
align_corners=self.align_corners)
|
| 56 |
+
ppm_outs.append(upsampled_ppm_out)
|
| 57 |
+
return ppm_outs
|
| 58 |
+
|
| 59 |
+
|
| 60 |
+
class BaseDecodeHead(nn.Module):
|
| 61 |
+
"""Base class for BaseDecodeHead.
|
| 62 |
+
|
| 63 |
+
Args:
|
| 64 |
+
in_channels (int|Sequence[int]): Input channels.
|
| 65 |
+
channels (int): Channels after modules, before conv_seg.
|
| 66 |
+
num_classes (int): Number of classes.
|
| 67 |
+
dropout_ratio (float): Ratio of dropout layer. Default: 0.1.
|
| 68 |
+
conv_cfg (dict|None): Config of conv layers. Default: None.
|
| 69 |
+
norm_cfg (dict|None): Config of norm layers. Default: None.
|
| 70 |
+
act_cfg (dict): Config of activation layers.
|
| 71 |
+
Default: dict(type='ReLU')
|
| 72 |
+
in_index (int|Sequence[int]): Input feature index. Default: -1
|
| 73 |
+
input_transform (str|None): Transformation type of input features.
|
| 74 |
+
Options: 'resize_concat', 'multiple_select', None.
|
| 75 |
+
'resize_concat': Multiple feature maps will be resize to the
|
| 76 |
+
same size as first one and than concat together.
|
| 77 |
+
Usually used in FCN head of HRNet.
|
| 78 |
+
'multiple_select': Multiple feature maps will be bundle into
|
| 79 |
+
a list and passed into decode head.
|
| 80 |
+
None: Only one select feature map is allowed.
|
| 81 |
+
Default: None.
|
| 82 |
+
loss_decode (dict): Config of decode loss.
|
| 83 |
+
Default: dict(type='CrossEntropyLoss').
|
| 84 |
+
ignore_index (int | None): The label index to be ignored. When using
|
| 85 |
+
masked BCE loss, ignore_index should be set to None. Default: 255
|
| 86 |
+
sampler (dict|None): The config of segmentation map sampler.
|
| 87 |
+
Default: None.
|
| 88 |
+
align_corners (bool): align_corners argument of F.interpolate.
|
| 89 |
+
Default: False.
|
| 90 |
+
"""
|
| 91 |
+
|
| 92 |
+
def __init__(self,
|
| 93 |
+
in_channels,
|
| 94 |
+
channels,
|
| 95 |
+
*,
|
| 96 |
+
num_classes,
|
| 97 |
+
dropout_ratio=0.1,
|
| 98 |
+
conv_cfg=None,
|
| 99 |
+
norm_cfg=None,
|
| 100 |
+
act_cfg=dict(type='ReLU'),
|
| 101 |
+
in_index=-1,
|
| 102 |
+
input_transform=None,
|
| 103 |
+
loss_decode=None,
|
| 104 |
+
ignore_index=255,
|
| 105 |
+
sampler=None,
|
| 106 |
+
align_corners=False):
|
| 107 |
+
super(BaseDecodeHead, self).__init__()
|
| 108 |
+
self._init_inputs(in_channels, in_index, input_transform)
|
| 109 |
+
self.channels = channels
|
| 110 |
+
self.num_classes = num_classes
|
| 111 |
+
self.dropout_ratio = dropout_ratio
|
| 112 |
+
self.conv_cfg = conv_cfg
|
| 113 |
+
self.norm_cfg = norm_cfg
|
| 114 |
+
self.act_cfg = act_cfg
|
| 115 |
+
self.in_index = in_index
|
| 116 |
+
self.ignore_index = ignore_index
|
| 117 |
+
self.align_corners = align_corners
|
| 118 |
+
if dropout_ratio > 0:
|
| 119 |
+
self.dropout = nn.Dropout2d(dropout_ratio)
|
| 120 |
+
else:
|
| 121 |
+
self.dropout = None
|
| 122 |
+
self.fp16_enabled = False
|
| 123 |
+
|
| 124 |
+
def extra_repr(self):
|
| 125 |
+
"""Extra repr."""
|
| 126 |
+
s = f'input_transform={self.input_transform}, ' \
|
| 127 |
+
f'ignore_index={self.ignore_index}, ' \
|
| 128 |
+
f'align_corners={self.align_corners}'
|
| 129 |
+
return s
|
| 130 |
+
|
| 131 |
+
def _init_inputs(self, in_channels, in_index, input_transform):
|
| 132 |
+
"""Check and initialize input transforms.
|
| 133 |
+
|
| 134 |
+
The in_channels, in_index and input_transform must match.
|
| 135 |
+
Specifically, when input_transform is None, only single feature map
|
| 136 |
+
will be selected. So in_channels and in_index must be of type int.
|
| 137 |
+
When input_transform
|
| 138 |
+
|
| 139 |
+
Args:
|
| 140 |
+
in_channels (int|Sequence[int]): Input channels.
|
| 141 |
+
in_index (int|Sequence[int]): Input feature index.
|
| 142 |
+
input_transform (str|None): Transformation type of input features.
|
| 143 |
+
Options: 'resize_concat', 'multiple_select', None.
|
| 144 |
+
'resize_concat': Multiple feature maps will be resize to the
|
| 145 |
+
same size as first one and than concat together.
|
| 146 |
+
Usually used in FCN head of HRNet.
|
| 147 |
+
'multiple_select': Multiple feature maps will be bundle into
|
| 148 |
+
a list and passed into decode head.
|
| 149 |
+
None: Only one select feature map is allowed.
|
| 150 |
+
"""
|
| 151 |
+
|
| 152 |
+
if input_transform is not None:
|
| 153 |
+
assert input_transform in ['resize_concat', 'multiple_select']
|
| 154 |
+
self.input_transform = input_transform
|
| 155 |
+
self.in_index = in_index
|
| 156 |
+
if input_transform is not None:
|
| 157 |
+
assert isinstance(in_channels, (list, tuple))
|
| 158 |
+
assert isinstance(in_index, (list, tuple))
|
| 159 |
+
assert len(in_channels) == len(in_index)
|
| 160 |
+
if input_transform == 'resize_concat':
|
| 161 |
+
self.in_channels = sum(in_channels)
|
| 162 |
+
else:
|
| 163 |
+
self.in_channels = in_channels
|
| 164 |
+
else:
|
| 165 |
+
assert isinstance(in_channels, int)
|
| 166 |
+
assert isinstance(in_index, int)
|
| 167 |
+
self.in_channels = in_channels
|
| 168 |
+
|
| 169 |
+
def init_weights(self):
|
| 170 |
+
"""Initialize weights of classification layer."""
|
| 171 |
+
# normal_init(self.conv_seg, mean=0, std=0.01)
|
| 172 |
+
# normal_init(self.conv1, mean=0, std=0.01)
|
| 173 |
+
|
| 174 |
+
def _transform_inputs(self, inputs):
|
| 175 |
+
"""Transform inputs for decoder.
|
| 176 |
+
|
| 177 |
+
Args:
|
| 178 |
+
inputs (list[Tensor]): List of multi-level img features.
|
| 179 |
+
|
| 180 |
+
Returns:
|
| 181 |
+
Tensor: The transformed inputs
|
| 182 |
+
"""
|
| 183 |
+
|
| 184 |
+
if self.input_transform == 'resize_concat':
|
| 185 |
+
inputs = [inputs[i] for i in self.in_index]
|
| 186 |
+
upsampled_inputs = [
|
| 187 |
+
resize(
|
| 188 |
+
input=x,
|
| 189 |
+
size=inputs[0].shape[2:],
|
| 190 |
+
mode='bilinear',
|
| 191 |
+
align_corners=self.align_corners) for x in inputs
|
| 192 |
+
]
|
| 193 |
+
inputs = torch.cat(upsampled_inputs, dim=1)
|
| 194 |
+
elif self.input_transform == 'multiple_select':
|
| 195 |
+
inputs = [inputs[i] for i in self.in_index]
|
| 196 |
+
else:
|
| 197 |
+
inputs = inputs[self.in_index]
|
| 198 |
+
|
| 199 |
+
return inputs
|
| 200 |
+
|
| 201 |
+
def forward(self, inputs):
|
| 202 |
+
"""Placeholder of forward function."""
|
| 203 |
+
pass
|
| 204 |
+
|
| 205 |
+
def forward_train(self, inputs, img_metas, gt_semantic_seg, train_cfg):
|
| 206 |
+
"""Forward function for training.
|
| 207 |
+
Args:
|
| 208 |
+
inputs (list[Tensor]): List of multi-level img features.
|
| 209 |
+
img_metas (list[dict]): List of image info dict where each dict
|
| 210 |
+
has: 'img_shape', 'scale_factor', 'flip', and may also contain
|
| 211 |
+
'filename', 'ori_shape', 'pad_shape', and 'img_norm_cfg'.
|
| 212 |
+
For details on the values of these keys see
|
| 213 |
+
`mmseg/datasets/pipelines/formatting.py:Collect`.
|
| 214 |
+
gt_semantic_seg (Tensor): Semantic segmentation masks
|
| 215 |
+
used if the architecture supports semantic segmentation task.
|
| 216 |
+
train_cfg (dict): The training config.
|
| 217 |
+
|
| 218 |
+
Returns:
|
| 219 |
+
dict[str, Tensor]: a dictionary of loss components
|
| 220 |
+
"""
|
| 221 |
+
seg_logits = self.forward(inputs)
|
| 222 |
+
losses = self.losses(seg_logits, gt_semantic_seg)
|
| 223 |
+
return losses
|
| 224 |
+
|
| 225 |
+
def forward_test(self, inputs, img_metas, test_cfg):
|
| 226 |
+
"""Forward function for testing.
|
| 227 |
+
|
| 228 |
+
Args:
|
| 229 |
+
inputs (list[Tensor]): List of multi-level img features.
|
| 230 |
+
img_metas (list[dict]): List of image info dict where each dict
|
| 231 |
+
has: 'img_shape', 'scale_factor', 'flip', and may also contain
|
| 232 |
+
'filename', 'ori_shape', 'pad_shape', and 'img_norm_cfg'.
|
| 233 |
+
For details on the values of these keys see
|
| 234 |
+
`mmseg/datasets/pipelines/formatting.py:Collect`.
|
| 235 |
+
test_cfg (dict): The testing config.
|
| 236 |
+
|
| 237 |
+
Returns:
|
| 238 |
+
Tensor: Output segmentation map.
|
| 239 |
+
"""
|
| 240 |
+
return self.forward(inputs)
|
| 241 |
+
|
| 242 |
+
|
| 243 |
+
class UPerHead(BaseDecodeHead):
|
| 244 |
+
def __init__(self, pool_scales=(1, 2, 3, 6), **kwargs):
|
| 245 |
+
super(UPerHead, self).__init__(
|
| 246 |
+
input_transform='multiple_select', **kwargs)
|
| 247 |
+
# FPN Module
|
| 248 |
+
self.lateral_convs = nn.ModuleList()
|
| 249 |
+
self.fpn_convs = nn.ModuleList()
|
| 250 |
+
for in_channels in self.in_channels: # skip the top layer
|
| 251 |
+
l_conv = ConvModule(
|
| 252 |
+
in_channels,
|
| 253 |
+
self.channels,
|
| 254 |
+
1,
|
| 255 |
+
conv_cfg=self.conv_cfg,
|
| 256 |
+
norm_cfg=self.norm_cfg,
|
| 257 |
+
act_cfg=self.act_cfg,
|
| 258 |
+
inplace=True)
|
| 259 |
+
fpn_conv = ConvModule(
|
| 260 |
+
self.channels,
|
| 261 |
+
self.channels,
|
| 262 |
+
3,
|
| 263 |
+
padding=1,
|
| 264 |
+
conv_cfg=self.conv_cfg,
|
| 265 |
+
norm_cfg=self.norm_cfg,
|
| 266 |
+
act_cfg=self.act_cfg,
|
| 267 |
+
inplace=True)
|
| 268 |
+
self.lateral_convs.append(l_conv)
|
| 269 |
+
self.fpn_convs.append(fpn_conv)
|
| 270 |
+
|
| 271 |
+
def forward(self, inputs):
|
| 272 |
+
"""Forward function."""
|
| 273 |
+
|
| 274 |
+
inputs = self._transform_inputs(inputs)
|
| 275 |
+
|
| 276 |
+
# build laterals
|
| 277 |
+
laterals = [
|
| 278 |
+
lateral_conv(inputs[i])
|
| 279 |
+
for i, lateral_conv in enumerate(self.lateral_convs)
|
| 280 |
+
]
|
| 281 |
+
|
| 282 |
+
# laterals.append(self.psp_forward(inputs))
|
| 283 |
+
|
| 284 |
+
# build top-down path
|
| 285 |
+
used_backbone_levels = len(laterals)
|
| 286 |
+
for i in range(used_backbone_levels - 1, 0, -1):
|
| 287 |
+
prev_shape = laterals[i - 1].shape[2:]
|
| 288 |
+
laterals[i - 1] += resize(
|
| 289 |
+
laterals[i],
|
| 290 |
+
size=prev_shape,
|
| 291 |
+
mode='bilinear',
|
| 292 |
+
align_corners=self.align_corners)
|
| 293 |
+
|
| 294 |
+
# build outputs
|
| 295 |
+
fpn_outs = [
|
| 296 |
+
self.fpn_convs[i](laterals[i])
|
| 297 |
+
for i in range(used_backbone_levels - 1)
|
| 298 |
+
]
|
| 299 |
+
# append psp feature
|
| 300 |
+
fpn_outs.append(laterals[-1])
|
| 301 |
+
|
| 302 |
+
return fpn_outs[0]
|
| 303 |
+
|
| 304 |
+
class PSP(BaseDecodeHead):
|
| 305 |
+
"""Unified Perceptual Parsing for Scene Understanding.
|
| 306 |
+
|
| 307 |
+
This head is the implementation of `UPerNet
|
| 308 |
+
<https://arxiv.org/abs/1807.10221>`_.
|
| 309 |
+
|
| 310 |
+
Args:
|
| 311 |
+
pool_scales (tuple[int]): Pooling scales used in Pooling Pyramid
|
| 312 |
+
Module applied on the last feature. Default: (1, 2, 3, 6).
|
| 313 |
+
"""
|
| 314 |
+
|
| 315 |
+
def __init__(self, pool_scales=(1, 2, 3, 6), **kwargs):
|
| 316 |
+
super(PSP, self).__init__(
|
| 317 |
+
input_transform='multiple_select', **kwargs)
|
| 318 |
+
# PSP Module
|
| 319 |
+
self.psp_modules = PPM(
|
| 320 |
+
pool_scales,
|
| 321 |
+
self.in_channels[-1],
|
| 322 |
+
self.channels,
|
| 323 |
+
conv_cfg=self.conv_cfg,
|
| 324 |
+
norm_cfg=self.norm_cfg,
|
| 325 |
+
act_cfg=self.act_cfg,
|
| 326 |
+
align_corners=self.align_corners)
|
| 327 |
+
self.bottleneck = ConvModule(
|
| 328 |
+
self.in_channels[-1] + len(pool_scales) * self.channels,
|
| 329 |
+
self.channels,
|
| 330 |
+
3,
|
| 331 |
+
padding=1,
|
| 332 |
+
conv_cfg=self.conv_cfg,
|
| 333 |
+
norm_cfg=self.norm_cfg,
|
| 334 |
+
act_cfg=self.act_cfg)
|
| 335 |
+
|
| 336 |
+
def psp_forward(self, inputs):
|
| 337 |
+
"""Forward function of PSP module."""
|
| 338 |
+
x = inputs[-1]
|
| 339 |
+
psp_outs = [x]
|
| 340 |
+
psp_outs.extend(self.psp_modules(x))
|
| 341 |
+
psp_outs = torch.cat(psp_outs, dim=1)
|
| 342 |
+
output = self.bottleneck(psp_outs)
|
| 343 |
+
|
| 344 |
+
return output
|
| 345 |
+
|
| 346 |
+
def forward(self, inputs):
|
| 347 |
+
"""Forward function."""
|
| 348 |
+
inputs = self._transform_inputs(inputs)
|
| 349 |
+
|
| 350 |
+
return self.psp_forward(inputs)
|
external/WildCamera/WildCamera/train/train_calibrator.py
ADDED
|
@@ -0,0 +1,297 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from __future__ import print_function, division
|
| 2 |
+
import os, sys, inspect, time, copy, warnings
|
| 3 |
+
project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
|
| 4 |
+
sys.path.insert(0, project_root)
|
| 5 |
+
warnings.filterwarnings('ignore')
|
| 6 |
+
|
| 7 |
+
import torch
|
| 8 |
+
import torch.backends.cudnn as cudnn
|
| 9 |
+
import torch.distributed as dist
|
| 10 |
+
import torch.multiprocessing as mp
|
| 11 |
+
from torch.utils.tensorboard import SummaryWriter
|
| 12 |
+
from torch.utils.data import ConcatDataset
|
| 13 |
+
from torchvision import transforms
|
| 14 |
+
|
| 15 |
+
import argparse
|
| 16 |
+
import numpy as np
|
| 17 |
+
from loguru import logger
|
| 18 |
+
from einops import rearrange
|
| 19 |
+
from pprint import pprint
|
| 20 |
+
|
| 21 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 22 |
+
from WildCamera.datasets.IncdDataset import IncdDataset
|
| 23 |
+
from tools.tools import to_cuda, intrinsic2incidence, resample_rgb, DistributedSamplerNoEvenlyDivisible, IncidenceLoss
|
| 24 |
+
from tools.visualization import tensor2disp, tensor2rgb
|
| 25 |
+
|
| 26 |
+
parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
|
| 27 |
+
|
| 28 |
+
parser.add_argument('--experiment_name', type=str, help='experiment name', default='multi-modality newcrfs')
|
| 29 |
+
parser.add_argument('--experiment_set', type=str, choices=['gsv', 'in_the_wild'], required=True)
|
| 30 |
+
parser.add_argument('--saving_location', type=str, help='saving location', default=None)
|
| 31 |
+
parser.add_argument('--encoder', type=str, help='type of encoder, base07, large07', default='large07')
|
| 32 |
+
parser.add_argument('--pretrain', type=str, help='path of pretrained encoder', default='model_zoo/swin_transformer/swin_large_patch4_window7_224_22k.pth')
|
| 33 |
+
parser.add_argument('--load_ckpt', type=str, help='path of ckpt', default=None)
|
| 34 |
+
parser.add_argument('--evaluation_only', action="store_true")
|
| 35 |
+
parser.add_argument('--l1_th', type=int, help='RANSAC threshold', default=0.02)
|
| 36 |
+
|
| 37 |
+
# Dataset
|
| 38 |
+
parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
|
| 39 |
+
parser.add_argument('--input_height', type=int, help='input height', default=480)
|
| 40 |
+
parser.add_argument('--input_width', type=int, help='input width', default=640)
|
| 41 |
+
|
| 42 |
+
# Training
|
| 43 |
+
parser.add_argument('--batch_size', type=int, help='batch size', default=16)
|
| 44 |
+
parser.add_argument('--loss', type=str, default='cosine', help='You can also choees l1 loss')
|
| 45 |
+
parser.add_argument('--steps_per_epoch', type=int, help='frequency for evaluation', default=1000)
|
| 46 |
+
parser.add_argument('--termination_epoch', type=int, help='epoch to stop training', default=25)
|
| 47 |
+
|
| 48 |
+
# Training Misc
|
| 49 |
+
parser.add_argument('--weight_decay', type=float, help='weight decay factor for optimization', default=1e-2)
|
| 50 |
+
parser.add_argument('--adam_eps', type=float, help='epsilon in Adam optimizer', default=1e-6)
|
| 51 |
+
parser.add_argument('--train_workers', type=int, help='dataloader workers', default=32)
|
| 52 |
+
parser.add_argument('--eval_workers', type=int, help='dataloader workers', default=2)
|
| 53 |
+
parser.add_argument('--learning_rate', type=float, help='initial learning rate', default=1e-5)
|
| 54 |
+
parser.add_argument('--end_learning_rate', type=float, help='end learning rate', default=-1)
|
| 55 |
+
|
| 56 |
+
# Augmentation
|
| 57 |
+
parser.add_argument('--dataset_favour_long', type=float, default=0.1, help='whether in training will see more samples from large dataset')
|
| 58 |
+
parser.add_argument('--augscale', type=float, default=2.0, help='The scale of Augmentation')
|
| 59 |
+
parser.add_argument('--no_change_prob', type=float, default=0.1, help='The probability of seeing original image')
|
| 60 |
+
parser.add_argument('--coloraugmentation', action="store_true")
|
| 61 |
+
parser.add_argument('--coloraugmentation_scale', type=float, default=0.0)
|
| 62 |
+
|
| 63 |
+
# Multi-gpu training
|
| 64 |
+
parser.add_argument('--gpu', type=int, help='GPU id to use.', default=None)
|
| 65 |
+
parser.add_argument('--dist_url', type=str, help='url used to set up distributed training', default='tcp://127.0.0.1:1235')
|
| 66 |
+
parser.add_argument('--dist_backend', type=str, help='distributed backend', default='nccl')
|
| 67 |
+
|
| 68 |
+
def main_worker(gpu, ngpus_per_node, args):
|
| 69 |
+
args.gpu = gpu
|
| 70 |
+
group = dist.init_process_group(backend=args.dist_backend, init_method=args.dist_url, world_size=args.world_size, rank=args.gpu)
|
| 71 |
+
|
| 72 |
+
# NeWCRFs model
|
| 73 |
+
model = NEWCRFIF(version=args.encoder, pretrained=args.pretrain)
|
| 74 |
+
model.train()
|
| 75 |
+
|
| 76 |
+
if args.load_ckpt is not None:
|
| 77 |
+
model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
|
| 78 |
+
model.eval()
|
| 79 |
+
logger.info("Load Model from %s" % args.load_ckpt)
|
| 80 |
+
|
| 81 |
+
if args.distributed:
|
| 82 |
+
if args.gpu is not None:
|
| 83 |
+
torch.cuda.set_device(args.gpu)
|
| 84 |
+
model.cuda(args.gpu)
|
| 85 |
+
args.batch_size = int(args.batch_size / ngpus_per_node)
|
| 86 |
+
args.train_workers = int(args.train_workers / ngpus_per_node)
|
| 87 |
+
model = torch.nn.parallel.DistributedDataParallel(model, device_ids=[args.gpu], find_unused_parameters=False)
|
| 88 |
+
else:
|
| 89 |
+
model.cuda()
|
| 90 |
+
model = torch.nn.parallel.DistributedDataParallel(model, find_unused_parameters=False)
|
| 91 |
+
else:
|
| 92 |
+
model = torch.nn.DataParallel(model)
|
| 93 |
+
model.cuda()
|
| 94 |
+
|
| 95 |
+
if args.distributed:
|
| 96 |
+
print("== Model Initialized on GPU: {}".format(args.gpu))
|
| 97 |
+
else:
|
| 98 |
+
print("== Model Initialized")
|
| 99 |
+
|
| 100 |
+
# Training parameters
|
| 101 |
+
optimizer = torch.optim.Adam([{'params': model.module.parameters()}], lr=args.learning_rate)
|
| 102 |
+
|
| 103 |
+
cudnn.benchmark = True
|
| 104 |
+
|
| 105 |
+
incidence_criterion = IncidenceLoss(loss=args.loss)
|
| 106 |
+
|
| 107 |
+
end_learning_rate = args.end_learning_rate if args.end_learning_rate != -1 else 0.1 * args.learning_rate
|
| 108 |
+
|
| 109 |
+
steps_per_epoch = args.steps_per_epoch
|
| 110 |
+
num_total_steps = 250000
|
| 111 |
+
epoch = int(num_total_steps / steps_per_epoch)
|
| 112 |
+
|
| 113 |
+
inv_normalize = transforms.Normalize(
|
| 114 |
+
mean=[-0.485 / 0.229, -0.456 / 0.224, -0.406 / 0.225],
|
| 115 |
+
std=[1 / 0.229, 1 / 0.224, 1 / 0.225]
|
| 116 |
+
)
|
| 117 |
+
|
| 118 |
+
if args.gpu == 0:
|
| 119 |
+
checkpoint_dir = os.path.join(args.saving_location, 'model_zoo', args.experiment_name)
|
| 120 |
+
writer = SummaryWriter(checkpoint_dir, flush_secs=30)
|
| 121 |
+
logger.add(os.path.join(checkpoint_dir, "{}.log".format(args.experiment_name)))
|
| 122 |
+
else:
|
| 123 |
+
writer = None
|
| 124 |
+
logger.remove()
|
| 125 |
+
logger.add(sys.stderr, level="ERROR")
|
| 126 |
+
|
| 127 |
+
if args.experiment_set == 'gsv':
|
| 128 |
+
from WildCamera.evaluation.evaluate_fov import EvaluateFov
|
| 129 |
+
evaluator = EvaluateFov()
|
| 130 |
+
elif args.experiment_set == 'in_the_wild':
|
| 131 |
+
from WildCamera.evaluation.evaluate_intrinsic import EvaluateIntrinsic
|
| 132 |
+
evaluator = EvaluateIntrinsic()
|
| 133 |
+
else:
|
| 134 |
+
raise NotImplementedError()
|
| 135 |
+
|
| 136 |
+
if args.evaluation_only:
|
| 137 |
+
evaluator.evaluate(
|
| 138 |
+
model,
|
| 139 |
+
args,
|
| 140 |
+
steps=0,
|
| 141 |
+
writer=writer,
|
| 142 |
+
group=group,
|
| 143 |
+
wtassumption=False,
|
| 144 |
+
)
|
| 145 |
+
evaluator.evaluate(
|
| 146 |
+
model,
|
| 147 |
+
args,
|
| 148 |
+
steps=0,
|
| 149 |
+
writer=writer,
|
| 150 |
+
group=group,
|
| 151 |
+
wtassumption=True,
|
| 152 |
+
)
|
| 153 |
+
return
|
| 154 |
+
|
| 155 |
+
for n_0 in range(epoch):
|
| 156 |
+
|
| 157 |
+
if n_0 > args.termination_epoch:
|
| 158 |
+
break
|
| 159 |
+
|
| 160 |
+
incdataset = IncdDataset(
|
| 161 |
+
data_root=args.data_path,
|
| 162 |
+
datasets_included=args.datasets_train,
|
| 163 |
+
ht=args.input_height,
|
| 164 |
+
wt=args.input_width,
|
| 165 |
+
augmentation=True,
|
| 166 |
+
split='train',
|
| 167 |
+
shuffleseed=int(n_0 + 1),
|
| 168 |
+
dataset_favour_long=args.dataset_favour_long,
|
| 169 |
+
augscale=args.augscale,
|
| 170 |
+
no_change_prob=args.no_change_prob,
|
| 171 |
+
coloraugmentation=args.coloraugmentation,
|
| 172 |
+
coloraugmentation_scale=args.coloraugmentation_scale
|
| 173 |
+
)
|
| 174 |
+
sampler = DistributedSamplerNoEvenlyDivisible(incdataset, shuffle=True)
|
| 175 |
+
dataloader = torch.utils.data.DataLoader(
|
| 176 |
+
incdataset,
|
| 177 |
+
sampler=sampler,
|
| 178 |
+
batch_size=int(args.batch_size),
|
| 179 |
+
num_workers=int(args.train_workers)
|
| 180 |
+
)
|
| 181 |
+
assert len(dataloader) > steps_per_epoch
|
| 182 |
+
dataloader = iter(dataloader)
|
| 183 |
+
sampler.set_epoch(n_0)
|
| 184 |
+
for global_step in range(int(n_0 * steps_per_epoch), int((n_0 + 1) * steps_per_epoch)):
|
| 185 |
+
sample_batched = next(dataloader)
|
| 186 |
+
sample_batched = to_cuda(sample_batched)
|
| 187 |
+
|
| 188 |
+
rgb, K = sample_batched['rgb'], sample_batched['K']
|
| 189 |
+
|
| 190 |
+
model.train()
|
| 191 |
+
incidence = model(rgb)
|
| 192 |
+
optimizer.zero_grad()
|
| 193 |
+
|
| 194 |
+
# Loss For Normal
|
| 195 |
+
loss = incidence_criterion(incidence, K)
|
| 196 |
+
loss.backward()
|
| 197 |
+
|
| 198 |
+
for param_group in optimizer.param_groups:
|
| 199 |
+
current_lr = (args.learning_rate - end_learning_rate) * (1 - global_step / num_total_steps) ** 0.9 + end_learning_rate
|
| 200 |
+
param_group['lr'] = current_lr
|
| 201 |
+
|
| 202 |
+
optimizer.step()
|
| 203 |
+
|
| 204 |
+
if np.mod(global_step, 1000) == 0 and args.gpu == 0:
|
| 205 |
+
b = 1
|
| 206 |
+
_, _, h, w = rgb.shape
|
| 207 |
+
|
| 208 |
+
rgb = inv_normalize(rgb)
|
| 209 |
+
vls1 = tensor2rgb(rgb, viewind=0)
|
| 210 |
+
|
| 211 |
+
device = rgb.device
|
| 212 |
+
incidence_gt = intrinsic2incidence(K[0:1], b, h, w, device)
|
| 213 |
+
incidence_gt = rearrange(incidence_gt.squeeze(dim=4), 'b h w d -> b d h w')
|
| 214 |
+
vls3 = tensor2rgb((incidence + 1) / 2, viewind=0)
|
| 215 |
+
vls4 = tensor2rgb((incidence_gt + 1) / 2, viewind=0)
|
| 216 |
+
|
| 217 |
+
vls = np.concatenate([np.array(vls1), np.array(vls3), np.array(vls4)], axis=0)
|
| 218 |
+
writer.add_image('visualization', (torch.from_numpy(vls).float() / 255).permute([2, 0, 1]), global_step)
|
| 219 |
+
|
| 220 |
+
if writer is not None and args.gpu == 0:
|
| 221 |
+
writer.add_scalar('loss/loss_incidence', loss.item(), global_step)
|
| 222 |
+
if np.mod(global_step, 500) == 0:
|
| 223 |
+
logger.info('Step %d, Epoch %d, loss %.3f' % (global_step, n_0, loss.item()))
|
| 224 |
+
|
| 225 |
+
evaluator.evaluate(
|
| 226 |
+
model,
|
| 227 |
+
args,
|
| 228 |
+
steps=global_step,
|
| 229 |
+
writer=writer,
|
| 230 |
+
group=group,
|
| 231 |
+
wtassumption=False,
|
| 232 |
+
)
|
| 233 |
+
|
| 234 |
+
evaluator.evaluate(
|
| 235 |
+
model,
|
| 236 |
+
args,
|
| 237 |
+
steps=global_step,
|
| 238 |
+
writer=writer,
|
| 239 |
+
group=group,
|
| 240 |
+
wtassumption=True,
|
| 241 |
+
)
|
| 242 |
+
|
| 243 |
+
def main():
|
| 244 |
+
args = parser.parse_args()
|
| 245 |
+
torch.cuda.empty_cache()
|
| 246 |
+
args.world_size = torch.cuda.device_count()
|
| 247 |
+
args.distributed = True
|
| 248 |
+
args.dist_url = 'tcp://127.0.0.1:' + str(np.random.randint(2000, 3000, 1).item())
|
| 249 |
+
|
| 250 |
+
if args.saving_location is None:
|
| 251 |
+
args.saving_location = project_root
|
| 252 |
+
|
| 253 |
+
if args.experiment_set == 'gsv':
|
| 254 |
+
args.datasets_train = [
|
| 255 |
+
'GSV'
|
| 256 |
+
]
|
| 257 |
+
args.datasets_eval = [
|
| 258 |
+
'GSV'
|
| 259 |
+
]
|
| 260 |
+
elif args.experiment_set == 'in_the_wild':
|
| 261 |
+
args.datasets_train = [
|
| 262 |
+
'Nuscenes',
|
| 263 |
+
'KITTI',
|
| 264 |
+
'Cityscapes',
|
| 265 |
+
'NYUv2',
|
| 266 |
+
'ARKitScenes',
|
| 267 |
+
'MegaDepth',
|
| 268 |
+
'SUN3D',
|
| 269 |
+
'MVImgNet',
|
| 270 |
+
'Objectron'
|
| 271 |
+
]
|
| 272 |
+
args.datasets_eval = [
|
| 273 |
+
'Nuscenes',
|
| 274 |
+
'KITTI',
|
| 275 |
+
'Cityscapes',
|
| 276 |
+
'NYUv2',
|
| 277 |
+
'ARKitScenes',
|
| 278 |
+
'MegaDepth',
|
| 279 |
+
'SUN3D',
|
| 280 |
+
'MVImgNet',
|
| 281 |
+
'Objectron',
|
| 282 |
+
'Waymo',
|
| 283 |
+
'BIWIRGBDID',
|
| 284 |
+
'RGBD',
|
| 285 |
+
'ScanNet',
|
| 286 |
+
'MVS',
|
| 287 |
+
'Scenes11'
|
| 288 |
+
]
|
| 289 |
+
|
| 290 |
+
else:
|
| 291 |
+
raise NotImplementedError()
|
| 292 |
+
|
| 293 |
+
pprint(vars(args))
|
| 294 |
+
mp.spawn(main_worker, nprocs=args.world_size, args=(args.world_size, args))
|
| 295 |
+
|
| 296 |
+
if __name__ == '__main__':
|
| 297 |
+
main()
|
external/WildCamera/__pycache__/hubconf.cpython-310.pyc
ADDED
|
Binary file (720 Bytes). View file
|
|
|
external/WildCamera/asset/download_demo_images.sh
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
cd asset
|
| 2 |
+
wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/asset/dollyzoom.tar
|
| 3 |
+
tar -xvf dollyzoom.tar
|
| 4 |
+
rm dollyzoom.tar
|
| 5 |
+
wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/asset/images-from-github-wt-intrinsic.tar
|
| 6 |
+
tar -xvf images-from-github-wt-intrinsic.tar
|
| 7 |
+
rm images-from-github-wt-intrinsic.tar
|
| 8 |
+
cd ..
|
external/WildCamera/asset/download_wildcamera_checkpoint.sh
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
cd model_zoo
|
| 2 |
+
mkdir Release
|
| 3 |
+
cd Release
|
| 4 |
+
wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/wild_camera_all.pth?download=true
|
| 5 |
+
mv wild_camera_all.pth?download=true wild_camera_all.pth
|
| 6 |
+
wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/wild_camera_gsv.pth?download=true
|
| 7 |
+
mv wild_camera_gsv.pth?download=true wild_camera_gsv.pth
|
| 8 |
+
cd ..
|
| 9 |
+
mkdir swin_transformer
|
| 10 |
+
cd swin_transformer
|
| 11 |
+
wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/swin_large_patch4_window7_224_22k.pth?download=true
|
| 12 |
+
mv swin_large_patch4_window7_224_22k.pth?download=true swin_large_patch4_window7_224_22k.pth
|
| 13 |
+
cd ..
|
external/WildCamera/asset/download_wildcamera_dataset.sh
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
external/WildCamera/demo/demo_dollyzoom.py
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, copy, tqdm, glob, natsort, torch
|
| 2 |
+
import numpy as np
|
| 3 |
+
project_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
| 4 |
+
|
| 5 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 6 |
+
from PIL import Image, ImageDraw, ImageFont
|
| 7 |
+
|
| 8 |
+
def draw_focalbar(rgbaug, fest, minfocal, maxfocal):
|
| 9 |
+
tmp = copy.deepcopy(rgbaug)
|
| 10 |
+
|
| 11 |
+
font_size = 28
|
| 12 |
+
font = ImageFont.truetype("arial.ttf", size=font_size)
|
| 13 |
+
|
| 14 |
+
rgbaug = copy.deepcopy(tmp)
|
| 15 |
+
w, h = rgbaug.size
|
| 16 |
+
rgbaug.putalpha(255)
|
| 17 |
+
|
| 18 |
+
paddingr = 0.1
|
| 19 |
+
barsty = h * paddingr
|
| 20 |
+
baredy = h * (1 - paddingr)
|
| 21 |
+
barx = w * 0.9
|
| 22 |
+
horizonbarnum = 7 * 5
|
| 23 |
+
horizonbarlen = 0.01
|
| 24 |
+
|
| 25 |
+
white = Image.new('RGBA', rgbaug.size, (255, 255, 255, 0))
|
| 26 |
+
draw = ImageDraw.Draw(white)
|
| 27 |
+
draw.rectangle(
|
| 28 |
+
[
|
| 29 |
+
(w * 0.9 - w * horizonbarlen - 15 / 640 * w, h * paddingr - 15 / 480 * h),
|
| 30 |
+
w * 0.9 + w * horizonbarlen + 35 / 640 * w, h * (1 - paddingr) + 15 / 480 * h
|
| 31 |
+
],
|
| 32 |
+
fill=(255, 255, 255, 128)
|
| 33 |
+
)
|
| 34 |
+
rgbaug = Image.alpha_composite(rgbaug, white)
|
| 35 |
+
|
| 36 |
+
draw = ImageDraw.Draw(rgbaug)
|
| 37 |
+
draw.line((barx, barsty, barx, baredy), fill=(0, 0, 0, 255), width=5)
|
| 38 |
+
|
| 39 |
+
for i in range(horizonbarnum + 1):
|
| 40 |
+
r = i / horizonbarnum
|
| 41 |
+
bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
|
| 42 |
+
barstx = w * 0.9 - w * horizonbarlen
|
| 43 |
+
baredx = w * 0.9 + w * horizonbarlen
|
| 44 |
+
draw.line((barstx, bary, baredx, bary), fill=(0, 0, 0, 255), width=int(3))
|
| 45 |
+
|
| 46 |
+
horizonbarnum = 7
|
| 47 |
+
horizonbarlen = 0.02
|
| 48 |
+
|
| 49 |
+
for i in range(horizonbarnum + 1):
|
| 50 |
+
r = i / horizonbarnum
|
| 51 |
+
bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
|
| 52 |
+
barstx = w * 0.9 - w * horizonbarlen
|
| 53 |
+
baredx = w * 0.9 + w * horizonbarlen
|
| 54 |
+
draw.line((barstx, bary, baredx, bary), fill=(0, 0, 0, 255), width=3)
|
| 55 |
+
|
| 56 |
+
textpadding = 30 / 640 * w
|
| 57 |
+
draw.text((barstx + textpadding, bary - 5 / 480 * h), str(int(maxfocal * r + minfocal * (1 - r))), fill=(0, 0, 0, 255), font=font)
|
| 58 |
+
|
| 59 |
+
r = (fest - minfocal) / (maxfocal - minfocal)
|
| 60 |
+
bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
|
| 61 |
+
barstx = w * 0.9 - w * horizonbarlen
|
| 62 |
+
baredx = w * 0.9 + w * horizonbarlen
|
| 63 |
+
draw.line((barstx, bary, baredx, bary), fill=(255, 165, 0, 255), width=5)
|
| 64 |
+
return rgbaug
|
| 65 |
+
|
| 66 |
+
def smooth_width(fests_precomputed, idx, width=3):
|
| 67 |
+
# Smooth as an average of Three Frames
|
| 68 |
+
stidx = idx - int((width - 1) / 2)
|
| 69 |
+
edidx = idx + int((width - 1) / 2)
|
| 70 |
+
|
| 71 |
+
if stidx < 0:
|
| 72 |
+
stidx = 0
|
| 73 |
+
|
| 74 |
+
if edidx >= len(fests_precomputed):
|
| 75 |
+
edidx = len(fests_precomputed) - 1
|
| 76 |
+
|
| 77 |
+
return np.mean(fests_precomputed[stidx:edidx+1])
|
| 78 |
+
|
| 79 |
+
|
| 80 |
+
@torch.no_grad()
|
| 81 |
+
def draw_dollyzoom(model):
|
| 82 |
+
""" In Validation, random sample N Images to mimic a real test set situation """
|
| 83 |
+
model.eval()
|
| 84 |
+
|
| 85 |
+
dollyzoom_folder = os.path.join(project_dir, 'asset', 'dollyzoom')
|
| 86 |
+
output_dollyzoom = os.path.join(project_dir, 'output', 'dollyzoom')
|
| 87 |
+
dollyzoomvideos = [
|
| 88 |
+
'dz1',
|
| 89 |
+
'dz2',
|
| 90 |
+
'dz3'
|
| 91 |
+
]
|
| 92 |
+
|
| 93 |
+
for dollyzoomvideo in dollyzoomvideos:
|
| 94 |
+
dollyzoomimg_folder = os.path.join(dollyzoom_folder, dollyzoomvideo)
|
| 95 |
+
jpgs = glob.glob(os.path.join(dollyzoomimg_folder, '*.jpg'))
|
| 96 |
+
jpgs = natsort.natsorted(jpgs)
|
| 97 |
+
|
| 98 |
+
fests = list()
|
| 99 |
+
for jpg in tqdm.tqdm(jpgs):
|
| 100 |
+
intrinsic, _ = model.inference(Image.open(jpg), wtassumption=False)
|
| 101 |
+
fest = intrinsic[0, 0]
|
| 102 |
+
fests.append(fest)
|
| 103 |
+
|
| 104 |
+
fests = np.array(fests)
|
| 105 |
+
minfocal = fests.min()
|
| 106 |
+
maxfocal = fests.max()
|
| 107 |
+
|
| 108 |
+
focalpadding = 0.1 * minfocal
|
| 109 |
+
|
| 110 |
+
jpgs_focalbar = list()
|
| 111 |
+
for idx, jpg in enumerate(jpgs):
|
| 112 |
+
fest = smooth_width(fests, idx, width=3)
|
| 113 |
+
jpgs_focalbar.append(draw_focalbar(Image.open(jpg), fest, minfocal=int(minfocal - focalpadding), maxfocal=int(maxfocal + focalpadding)))
|
| 114 |
+
|
| 115 |
+
output_folder = os.path.join(output_dollyzoom, dollyzoomvideo)
|
| 116 |
+
os.makedirs(output_folder, exist_ok=True)
|
| 117 |
+
for idx, jpgimgage in enumerate(jpgs_focalbar):
|
| 118 |
+
jpg_path = os.path.join(output_folder, '{}.png'.format(str(idx)))
|
| 119 |
+
jpgimgage.save(jpg_path)
|
| 120 |
+
|
| 121 |
+
return
|
| 122 |
+
|
| 123 |
+
if __name__ == '__main__':
|
| 124 |
+
# NeWCRFs model
|
| 125 |
+
model = NEWCRFIF(version='large07', pretrained=None)
|
| 126 |
+
model.eval()
|
| 127 |
+
model.cuda()
|
| 128 |
+
|
| 129 |
+
script_dir = os.path.dirname(os.path.realpath(__file__))
|
| 130 |
+
ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
|
| 131 |
+
model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
|
| 132 |
+
|
| 133 |
+
draw_dollyzoom(model)
|
external/WildCamera/demo/demo_inference.py
ADDED
|
@@ -0,0 +1,30 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, torch
|
| 2 |
+
project_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
|
| 3 |
+
|
| 4 |
+
from PIL import Image
|
| 5 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
if __name__ == '__main__':
|
| 9 |
+
# NeWCRFs model
|
| 10 |
+
model = NEWCRFIF(version='large07', pretrained=None)
|
| 11 |
+
model.eval()
|
| 12 |
+
model.cuda()
|
| 13 |
+
|
| 14 |
+
script_dir = os.path.dirname(os.path.realpath(__file__))
|
| 15 |
+
ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
|
| 16 |
+
model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
|
| 17 |
+
|
| 18 |
+
images_folder = os.path.join(project_dir, 'asset', 'images-from-github-wt-intrinsic')
|
| 19 |
+
info_path = os.path.join(images_folder, 'intrinsic_gt.txt')
|
| 20 |
+
with open(info_path) as file:
|
| 21 |
+
infos = [line.rstrip() for line in file]
|
| 22 |
+
|
| 23 |
+
for idx, info in enumerate(infos):
|
| 24 |
+
imgname, focalgt, source = info.split(' ')
|
| 25 |
+
|
| 26 |
+
images_path = os.path.join(images_folder, imgname)
|
| 27 |
+
intrinsic, _ = model.inference(Image.open(images_path), wtassumption=False)
|
| 28 |
+
focal = intrinsic[0, 0].item()
|
| 29 |
+
|
| 30 |
+
print("Image Name: %s, Est Focal %.1f, Gt Focal %.1f, Source - %s" % (imgname, focal, float(focalgt), source))
|
external/WildCamera/demo/demo_restoration.py
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os, sys, inspect
|
| 2 |
+
project_root = os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))))
|
| 3 |
+
sys.path.insert(0, project_root)
|
| 4 |
+
|
| 5 |
+
import PIL.Image as Image
|
| 6 |
+
import torch
|
| 7 |
+
from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
|
| 8 |
+
|
| 9 |
+
@torch.no_grad()
|
| 10 |
+
def main():
|
| 11 |
+
model = NEWCRFIF(version='large07', pretrained=None)
|
| 12 |
+
model.eval()
|
| 13 |
+
model.cuda()
|
| 14 |
+
|
| 15 |
+
script_dir = os.path.dirname(os.path.realpath(__file__))
|
| 16 |
+
ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
|
| 17 |
+
model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
|
| 18 |
+
|
| 19 |
+
img_folder = os.path.join(project_root, 'asset', 'image_restoration')
|
| 20 |
+
img_names = [
|
| 21 |
+
'lenna_distorted_cropright.jpg',
|
| 22 |
+
'lenna_distorted_cropleft.jpg',
|
| 23 |
+
]
|
| 24 |
+
for img_name in img_names:
|
| 25 |
+
img_path = os.path.join(img_folder, img_name)
|
| 26 |
+
img = Image.open(img_path)
|
| 27 |
+
|
| 28 |
+
export_root = os.path.join(img_folder, '{}_restored.jpg'.format(img_name.split('.')[0]))
|
| 29 |
+
intrinsic, _ = model.inference(img, wtassumption=False)
|
| 30 |
+
model.restore_image(img, intrinsic).save(export_root)
|
| 31 |
+
|
| 32 |
+
if __name__ == '__main__':
|
| 33 |
+
main()
|
external/WildCamera/splits/arkitscenes_test.txt
ADDED
|
@@ -0,0 +1,800 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ARKitScenes_test Validation_41069021_401.371_00005806
|
| 2 |
+
ARKitScenes_test Validation_41069021_412.367_00006466
|
| 3 |
+
ARKitScenes_test Validation_41069021_417.365_00006766
|
| 4 |
+
ARKitScenes_test Validation_41069021_421.363_00007006
|
| 5 |
+
ARKitScenes_test Validation_41069021_423.362_00007126
|
| 6 |
+
ARKitScenes_test Validation_41069021_435.357_00007846
|
| 7 |
+
ARKitScenes_test Validation_41069021_450.368_00008747
|
| 8 |
+
ARKitScenes_test Validation_41069025_624.863_00002137
|
| 9 |
+
ARKitScenes_test Validation_41069025_648.870_00003578
|
| 10 |
+
ARKitScenes_test Validation_41069025_731.869_00008560
|
| 11 |
+
ARKitScenes_test Validation_41069042_3044.239_00000085
|
| 12 |
+
ARKitScenes_test Validation_41069042_3045.222_00000144
|
| 13 |
+
ARKitScenes_test Validation_41069042_3063.231_00001225
|
| 14 |
+
ARKitScenes_test Validation_41069042_3070.228_00001645
|
| 15 |
+
ARKitScenes_test Validation_41069042_3073.227_00001825
|
| 16 |
+
ARKitScenes_test Validation_41069043_2820.631_00001297
|
| 17 |
+
ARKitScenes_test Validation_41069043_2826.629_00001657
|
| 18 |
+
ARKitScenes_test Validation_41069043_2830.627_00001897
|
| 19 |
+
ARKitScenes_test Validation_41069043_2892.635_00005619
|
| 20 |
+
ARKitScenes_test Validation_41069046_2929.336_00001117
|
| 21 |
+
ARKitScenes_test Validation_41069046_2988.328_00004658
|
| 22 |
+
ARKitScenes_test Validation_41069046_2989.328_00004718
|
| 23 |
+
ARKitScenes_test Validation_41069046_2997.325_00005198
|
| 24 |
+
ARKitScenes_test Validation_41069046_3024.330_00006819
|
| 25 |
+
ARKitScenes_test Validation_41069051_5085.217_00000211
|
| 26 |
+
ARKitScenes_test Validation_41142278_4003.081_00000097
|
| 27 |
+
ARKitScenes_test Validation_41142278_4070.070_00004118
|
| 28 |
+
ARKitScenes_test Validation_41142280_3837.982_00000937
|
| 29 |
+
ARKitScenes_test Validation_41142281_3964.380_00002798
|
| 30 |
+
ARKitScenes_test Validation_42444946_223634.869_00001657
|
| 31 |
+
ARKitScenes_test Validation_42444946_223654.878_00002858
|
| 32 |
+
ARKitScenes_test Validation_42444946_223693.879_00005199
|
| 33 |
+
ARKitScenes_test Validation_42444950_223834.571_00002198
|
| 34 |
+
ARKitScenes_test Validation_42444950_223862.577_00003879
|
| 35 |
+
ARKitScenes_test Validation_42444966_40376.779_00002863
|
| 36 |
+
ARKitScenes_test Validation_42444966_40381.794_00003164
|
| 37 |
+
ARKitScenes_test Validation_42444966_40382.793_00003224
|
| 38 |
+
ARKitScenes_test Validation_42444966_40391.790_00003764
|
| 39 |
+
ARKitScenes_test Validation_42444966_40403.785_00004484
|
| 40 |
+
ARKitScenes_test Validation_42444968_40282.385_00002269
|
| 41 |
+
ARKitScenes_test Validation_42444968_40298.395_00003230
|
| 42 |
+
ARKitScenes_test Validation_42444968_40314.388_00004190
|
| 43 |
+
ARKitScenes_test Validation_42444968_40317.387_00004370
|
| 44 |
+
ARKitScenes_test Validation_42444976_40167.383_00001597
|
| 45 |
+
ARKitScenes_test Validation_42444976_40181.394_00002438
|
| 46 |
+
ARKitScenes_test Validation_42444976_40230.390_00005379
|
| 47 |
+
ARKitScenes_test Validation_42445021_48888.256_00000516
|
| 48 |
+
ARKitScenes_test Validation_42445026_48955.063_00001189
|
| 49 |
+
ARKitScenes_test Validation_42445031_49451.392_00000403
|
| 50 |
+
ARKitScenes_test Validation_42445031_49460.389_00000943
|
| 51 |
+
ARKitScenes_test Validation_42445031_49507.387_00003764
|
| 52 |
+
ARKitScenes_test Validation_42445031_49509.403_00003885
|
| 53 |
+
ARKitScenes_test Validation_42445031_49515.400_00004245
|
| 54 |
+
ARKitScenes_test Validation_42445429_50317.089_00000037
|
| 55 |
+
ARKitScenes_test Validation_42445441_50215.779_00000096
|
| 56 |
+
ARKitScenes_test Validation_42445441_50225.792_00000697
|
| 57 |
+
ARKitScenes_test Validation_42445448_50409.603_00000056
|
| 58 |
+
ARKitScenes_test Validation_42445448_50484.590_00004557
|
| 59 |
+
ARKitScenes_test Validation_42445448_50505.581_00005817
|
| 60 |
+
ARKitScenes_test Validation_42445991_75294.425_00002323
|
| 61 |
+
ARKitScenes_test Validation_42445991_75309.436_00003224
|
| 62 |
+
ARKitScenes_test Validation_42445991_75325.429_00004184
|
| 63 |
+
ARKitScenes_test Validation_42446038_347874.972_00000943
|
| 64 |
+
ARKitScenes_test Validation_42446049_347782.076_00000517
|
| 65 |
+
ARKitScenes_test Validation_42446100_77922.291_00000937
|
| 66 |
+
ARKitScenes_test Validation_42446100_77925.290_00001117
|
| 67 |
+
ARKitScenes_test Validation_42446100_77951.280_00002677
|
| 68 |
+
ARKitScenes_test Validation_42446100_77954.295_00002858
|
| 69 |
+
ARKitScenes_test Validation_42446100_77956.294_00002978
|
| 70 |
+
ARKitScenes_test Validation_42446100_77967.290_00003638
|
| 71 |
+
ARKitScenes_test Validation_42446100_77971.288_00003878
|
| 72 |
+
ARKitScenes_test Validation_42446100_77978.285_00004298
|
| 73 |
+
ARKitScenes_test Validation_42446100_78000.293_00005619
|
| 74 |
+
ARKitScenes_test Validation_42446100_78004.292_00005859
|
| 75 |
+
ARKitScenes_test Validation_42446100_78012.289_00006339
|
| 76 |
+
ARKitScenes_test Validation_42446100_78018.286_00006699
|
| 77 |
+
ARKitScenes_test Validation_42446100_78021.285_00006879
|
| 78 |
+
ARKitScenes_test Validation_42446100_78025.284_00007119
|
| 79 |
+
ARKitScenes_test Validation_42446103_78053.189_00000043
|
| 80 |
+
ARKitScenes_test Validation_42446103_78059.187_00000403
|
| 81 |
+
ARKitScenes_test Validation_42446103_78098.188_00002744
|
| 82 |
+
ARKitScenes_test Validation_42446103_78113.182_00003644
|
| 83 |
+
ARKitScenes_test Validation_42446103_78120.196_00004065
|
| 84 |
+
ARKitScenes_test Validation_42446103_78124.194_00004305
|
| 85 |
+
ARKitScenes_test Validation_42446103_78125.194_00004365
|
| 86 |
+
ARKitScenes_test Validation_42446103_78131.192_00004725
|
| 87 |
+
ARKitScenes_test Validation_42446103_78165.195_00006766
|
| 88 |
+
ARKitScenes_test Validation_42446103_78170.193_00007066
|
| 89 |
+
ARKitScenes_test Validation_42446103_78176.191_00007426
|
| 90 |
+
ARKitScenes_test Validation_42446114_78209.994_00001298
|
| 91 |
+
ARKitScenes_test Validation_42446114_78218.990_00001838
|
| 92 |
+
ARKitScenes_test Validation_42446114_78220.989_00001958
|
| 93 |
+
ARKitScenes_test Validation_42446114_78221.989_00002018
|
| 94 |
+
ARKitScenes_test Validation_42446114_78238.982_00003038
|
| 95 |
+
ARKitScenes_test Validation_42446114_78241.981_00003218
|
| 96 |
+
ARKitScenes_test Validation_42446114_78258.991_00004239
|
| 97 |
+
ARKitScenes_test Validation_42446114_78289.995_00006100
|
| 98 |
+
ARKitScenes_test Validation_42446114_78314.986_00007600
|
| 99 |
+
ARKitScenes_test Validation_42446114_78322.982_00008080
|
| 100 |
+
ARKitScenes_test Validation_42446114_78327.997_00008381
|
| 101 |
+
ARKitScenes_test Validation_42446156_79145.119_00000577
|
| 102 |
+
ARKitScenes_test Validation_42446163_79567.523_00000097
|
| 103 |
+
ARKitScenes_test Validation_42446163_79575.520_00000577
|
| 104 |
+
ARKitScenes_test Validation_42446165_79787.719_00004358
|
| 105 |
+
ARKitScenes_test Validation_42446167_79666.734_00001424
|
| 106 |
+
ARKitScenes_test Validation_42446167_79674.731_00001904
|
| 107 |
+
ARKitScenes_test Validation_42446167_79675.730_00001964
|
| 108 |
+
ARKitScenes_test Validation_42446167_79693.723_00003044
|
| 109 |
+
ARKitScenes_test Validation_42446167_79708.734_00003945
|
| 110 |
+
ARKitScenes_test Validation_42446517_197446.270_00001717
|
| 111 |
+
ARKitScenes_test Validation_42446517_197455.283_00002258
|
| 112 |
+
ARKitScenes_test Validation_42446519_197493.884_00000578
|
| 113 |
+
ARKitScenes_test Validation_42446519_197515.875_00001898
|
| 114 |
+
ARKitScenes_test Validation_42446519_197554.875_00004239
|
| 115 |
+
ARKitScenes_test Validation_42446519_197555.875_00004299
|
| 116 |
+
ARKitScenes_test Validation_42446519_197593.876_00006580
|
| 117 |
+
ARKitScenes_test Validation_42446522_197357.873_00002647
|
| 118 |
+
ARKitScenes_test Validation_42446522_197389.877_00004568
|
| 119 |
+
ARKitScenes_test Validation_42446522_197392.875_00004748
|
| 120 |
+
ARKitScenes_test Validation_42446522_197394.874_00004868
|
| 121 |
+
ARKitScenes_test Validation_42446527_198630.850_00000301
|
| 122 |
+
ARKitScenes_test Validation_42446529_198723.845_00000133
|
| 123 |
+
ARKitScenes_test Validation_42446529_198742.837_00001273
|
| 124 |
+
ARKitScenes_test Validation_42446529_198770.842_00002954
|
| 125 |
+
ARKitScenes_test Validation_42446532_198850.942_00002600
|
| 126 |
+
ARKitScenes_test Validation_42446532_198871.950_00003861
|
| 127 |
+
ARKitScenes_test Validation_42446533_200402.294_00000535
|
| 128 |
+
ARKitScenes_test Validation_42446533_200404.294_00000655
|
| 129 |
+
ARKitScenes_test Validation_42446535_200530.492_00003188
|
| 130 |
+
ARKitScenes_test Validation_42446540_201462.348_00000691
|
| 131 |
+
ARKitScenes_test Validation_42446540_201472.344_00001291
|
| 132 |
+
ARKitScenes_test Validation_42446540_201478.342_00001651
|
| 133 |
+
ARKitScenes_test Validation_42446540_201482.340_00001891
|
| 134 |
+
ARKitScenes_test Validation_42446540_201489.337_00002311
|
| 135 |
+
ARKitScenes_test Validation_42446540_201499.350_00002912
|
| 136 |
+
ARKitScenes_test Validation_42446540_201502.348_00003092
|
| 137 |
+
ARKitScenes_test Validation_42446540_201517.342_00003992
|
| 138 |
+
ARKitScenes_test Validation_42446540_201524.339_00004412
|
| 139 |
+
ARKitScenes_test Validation_42446540_201526.338_00004532
|
| 140 |
+
ARKitScenes_test Validation_42446540_201534.335_00005012
|
| 141 |
+
ARKitScenes_test Validation_42446540_201536.351_00005133
|
| 142 |
+
ARKitScenes_test Validation_42446541_201563.440_00000943
|
| 143 |
+
ARKitScenes_test Validation_42446541_201567.438_00001183
|
| 144 |
+
ARKitScenes_test Validation_42446541_201628.446_00004845
|
| 145 |
+
ARKitScenes_test Validation_42446541_201636.443_00005325
|
| 146 |
+
ARKitScenes_test Validation_42446541_201638.442_00005445
|
| 147 |
+
ARKitScenes_test Validation_42446541_201657.434_00006585
|
| 148 |
+
ARKitScenes_test Validation_42446541_201661.433_00006825
|
| 149 |
+
ARKitScenes_test Validation_42897501_270032.593_00001466
|
| 150 |
+
ARKitScenes_test Validation_42897501_270048.586_00002426
|
| 151 |
+
ARKitScenes_test Validation_42897501_270050.586_00002546
|
| 152 |
+
ARKitScenes_test Validation_42897501_270053.584_00002726
|
| 153 |
+
ARKitScenes_test Validation_42897501_270069.578_00003686
|
| 154 |
+
ARKitScenes_test Validation_42897504_269846.486_00000259
|
| 155 |
+
ARKitScenes_test Validation_42897504_269883.487_00002480
|
| 156 |
+
ARKitScenes_test Validation_42897504_269889.485_00002840
|
| 157 |
+
ARKitScenes_test Validation_42897504_269902.479_00003620
|
| 158 |
+
ARKitScenes_test Validation_42897508_269962.088_00002384
|
| 159 |
+
ARKitScenes_test Validation_42897508_269975.083_00003164
|
| 160 |
+
ARKitScenes_test Validation_42897521_273988.900_00002162
|
| 161 |
+
ARKitScenes_test Validation_42897526_273865.400_00001255
|
| 162 |
+
ARKitScenes_test Validation_42897526_273873.397_00001735
|
| 163 |
+
ARKitScenes_test Validation_42897528_273919.095_00001141
|
| 164 |
+
ARKitScenes_test Validation_42897528_273940.103_00002402
|
| 165 |
+
ARKitScenes_test Validation_42897541_275396.527_00001285
|
| 166 |
+
ARKitScenes_test Validation_42897541_275404.540_00001766
|
| 167 |
+
ARKitScenes_test Validation_42897541_275436.527_00003686
|
| 168 |
+
ARKitScenes_test Validation_42897542_275468.131_00000811
|
| 169 |
+
ARKitScenes_test Validation_42897542_275483.141_00001712
|
| 170 |
+
ARKitScenes_test Validation_42897549_276508.368_00001387
|
| 171 |
+
ARKitScenes_test Validation_42897549_276512.366_00001627
|
| 172 |
+
ARKitScenes_test Validation_42897552_277361.813_00000217
|
| 173 |
+
ARKitScenes_test Validation_42897554_277333.008_00001135
|
| 174 |
+
ARKitScenes_test Validation_42897559_277434.600_00000241
|
| 175 |
+
ARKitScenes_test Validation_42897561_278930.716_00001159
|
| 176 |
+
ARKitScenes_test Validation_42897561_278945.710_00002059
|
| 177 |
+
ARKitScenes_test Validation_42897561_278946.710_00002119
|
| 178 |
+
ARKitScenes_test Validation_42897561_278950.708_00002359
|
| 179 |
+
ARKitScenes_test Validation_42897566_279057.014_00001076
|
| 180 |
+
ARKitScenes_test Validation_42897566_279079.005_00002396
|
| 181 |
+
ARKitScenes_test Validation_42897599_476331.200_00001789
|
| 182 |
+
ARKitScenes_test Validation_42897599_476342.212_00002450
|
| 183 |
+
ARKitScenes_test Validation_42897599_476351.209_00002990
|
| 184 |
+
ARKitScenes_test Validation_42897599_476408.202_00006411
|
| 185 |
+
ARKitScenes_test Validation_42897599_476442.205_00008452
|
| 186 |
+
ARKitScenes_test Validation_42897599_476474.208_00010373
|
| 187 |
+
ARKitScenes_test Validation_42897599_476495.200_00011633
|
| 188 |
+
ARKitScenes_test Validation_42897599_476498.199_00011813
|
| 189 |
+
ARKitScenes_test Validation_42897599_476518.207_00013014
|
| 190 |
+
ARKitScenes_test Validation_42897599_476522.205_00013254
|
| 191 |
+
ARKitScenes_test Validation_42897599_476559.207_00015475
|
| 192 |
+
ARKitScenes_test Validation_42897647_478140.187_00000643
|
| 193 |
+
ARKitScenes_test Validation_42897647_478179.188_00002984
|
| 194 |
+
ARKitScenes_test Validation_42897647_478188.184_00003524
|
| 195 |
+
ARKitScenes_test Validation_42897647_478200.179_00004244
|
| 196 |
+
ARKitScenes_test Validation_42897647_478203.178_00004424
|
| 197 |
+
ARKitScenes_test Validation_42897647_478207.176_00004664
|
| 198 |
+
ARKitScenes_test Validation_42897647_478215.190_00005145
|
| 199 |
+
ARKitScenes_test Validation_42897651_478324.278_00004796
|
| 200 |
+
ARKitScenes_test Validation_42897651_478325.278_00004856
|
| 201 |
+
ARKitScenes_test Validation_42897651_478334.274_00005396
|
| 202 |
+
ARKitScenes_test Validation_42897651_478344.286_00005997
|
| 203 |
+
ARKitScenes_test Validation_42897651_478354.282_00006597
|
| 204 |
+
ARKitScenes_test Validation_42897651_478357.281_00006777
|
| 205 |
+
ARKitScenes_test Validation_42897651_478370.276_00007557
|
| 206 |
+
ARKitScenes_test Validation_42897651_478433.283_00011339
|
| 207 |
+
ARKitScenes_test Validation_42897651_478439.281_00011699
|
| 208 |
+
ARKitScenes_test Validation_42897651_478440.280_00011759
|
| 209 |
+
ARKitScenes_test Validation_42897651_478458.273_00012839
|
| 210 |
+
ARKitScenes_test Validation_42897651_478463.271_00013139
|
| 211 |
+
ARKitScenes_test Validation_42897651_478473.283_00013740
|
| 212 |
+
ARKitScenes_test Validation_42897651_478477.282_00013980
|
| 213 |
+
ARKitScenes_test Validation_42897651_478481.280_00014220
|
| 214 |
+
ARKitScenes_test Validation_42897651_478482.280_00014280
|
| 215 |
+
ARKitScenes_test Validation_42897651_478498.273_00015240
|
| 216 |
+
ARKitScenes_test Validation_42897667_478043.077_00002797
|
| 217 |
+
ARKitScenes_test Validation_42897667_478052.090_00003338
|
| 218 |
+
ARKitScenes_test Validation_42897667_478080.078_00005018
|
| 219 |
+
ARKitScenes_test Validation_42897667_478084.077_00005258
|
| 220 |
+
ARKitScenes_test Validation_42897667_478120.079_00007419
|
| 221 |
+
ARKitScenes_test Validation_42897688_479152.078_00000516
|
| 222 |
+
ARKitScenes_test Validation_42897688_479155.076_00000696
|
| 223 |
+
ARKitScenes_test Validation_42897688_479181.082_00002257
|
| 224 |
+
ARKitScenes_test Validation_42897688_479189.079_00002737
|
| 225 |
+
ARKitScenes_test Validation_42897688_479227.080_00005018
|
| 226 |
+
ARKitScenes_test Validation_42897688_479273.078_00007779
|
| 227 |
+
ARKitScenes_test Validation_42897688_479276.077_00007959
|
| 228 |
+
ARKitScenes_test Validation_42897688_479302.082_00009520
|
| 229 |
+
ARKitScenes_test Validation_42897688_479303.082_00009580
|
| 230 |
+
ARKitScenes_test Validation_42898811_167554.386_00005018
|
| 231 |
+
ARKitScenes_test Validation_42898816_167393.902_00002444
|
| 232 |
+
ARKitScenes_test Validation_42898818_168178.187_00001598
|
| 233 |
+
ARKitScenes_test Validation_42898822_168207.574_00000157
|
| 234 |
+
ARKitScenes_test Validation_42898822_168249.574_00002678
|
| 235 |
+
ARKitScenes_test Validation_42898822_168262.585_00003459
|
| 236 |
+
ARKitScenes_test Validation_42898826_168099.186_00000103
|
| 237 |
+
ARKitScenes_test Validation_42898849_196451.916_00000397
|
| 238 |
+
ARKitScenes_test Validation_42898849_196462.928_00001058
|
| 239 |
+
ARKitScenes_test Validation_42898849_196469.926_00001478
|
| 240 |
+
ARKitScenes_test Validation_42898849_196475.923_00001838
|
| 241 |
+
ARKitScenes_test Validation_42898849_196483.920_00002318
|
| 242 |
+
ARKitScenes_test Validation_42898849_196484.920_00002378
|
| 243 |
+
ARKitScenes_test Validation_42898849_196497.915_00003158
|
| 244 |
+
ARKitScenes_test Validation_42898849_196517.923_00004359
|
| 245 |
+
ARKitScenes_test Validation_42898849_196543.913_00005919
|
| 246 |
+
ARKitScenes_test Validation_42898849_196544.913_00005979
|
| 247 |
+
ARKitScenes_test Validation_42898849_196546.928_00006100
|
| 248 |
+
ARKitScenes_test Validation_42898849_196563.922_00007120
|
| 249 |
+
ARKitScenes_test Validation_42898849_196582.914_00008260
|
| 250 |
+
ARKitScenes_test Validation_42898854_196214.727_00001928
|
| 251 |
+
ARKitScenes_test Validation_42898854_196218.725_00002168
|
| 252 |
+
ARKitScenes_test Validation_42898854_196260.725_00004689
|
| 253 |
+
ARKitScenes_test Validation_42898854_196276.719_00005649
|
| 254 |
+
ARKitScenes_test Validation_42898854_196281.717_00005949
|
| 255 |
+
ARKitScenes_test Validation_42898854_196289.714_00006429
|
| 256 |
+
ARKitScenes_test Validation_42898854_196292.712_00006609
|
| 257 |
+
ARKitScenes_test Validation_42898854_196297.727_00006910
|
| 258 |
+
ARKitScenes_test Validation_42898862_196317.719_00000391
|
| 259 |
+
ARKitScenes_test Validation_42898862_196324.716_00000811
|
| 260 |
+
ARKitScenes_test Validation_42898862_196328.715_00001051
|
| 261 |
+
ARKitScenes_test Validation_42898862_196330.714_00001171
|
| 262 |
+
ARKitScenes_test Validation_42898862_196343.726_00001952
|
| 263 |
+
ARKitScenes_test Validation_42898862_196348.724_00002252
|
| 264 |
+
ARKitScenes_test Validation_42898862_196352.722_00002492
|
| 265 |
+
ARKitScenes_test Validation_42898862_196361.718_00003032
|
| 266 |
+
ARKitScenes_test Validation_42898862_196364.717_00003212
|
| 267 |
+
ARKitScenes_test Validation_42898862_196371.714_00003632
|
| 268 |
+
ARKitScenes_test Validation_42898862_196382.727_00004293
|
| 269 |
+
ARKitScenes_test Validation_42898862_196385.726_00004473
|
| 270 |
+
ARKitScenes_test Validation_42898862_196387.725_00004593
|
| 271 |
+
ARKitScenes_test Validation_42898862_196396.721_00005133
|
| 272 |
+
ARKitScenes_test Validation_42898862_196418.713_00006453
|
| 273 |
+
ARKitScenes_test Validation_42898862_196423.727_00006754
|
| 274 |
+
ARKitScenes_test Validation_42899461_195801.429_00000997
|
| 275 |
+
ARKitScenes_test Validation_42899461_195803.428_00001117
|
| 276 |
+
ARKitScenes_test Validation_42899461_195807.426_00001357
|
| 277 |
+
ARKitScenes_test Validation_42899461_195865.419_00004838
|
| 278 |
+
ARKitScenes_test Validation_42899461_195869.434_00005079
|
| 279 |
+
ARKitScenes_test Validation_42899461_195881.429_00005799
|
| 280 |
+
ARKitScenes_test Validation_42899471_195673.532_00000938
|
| 281 |
+
ARKitScenes_test Validation_42899471_195705.535_00002859
|
| 282 |
+
ARKitScenes_test Validation_42899471_195706.535_00002919
|
| 283 |
+
ARKitScenes_test Validation_42899471_195710.533_00003159
|
| 284 |
+
ARKitScenes_test Validation_42899611_228365.713_00000997
|
| 285 |
+
ARKitScenes_test Validation_42899611_228379.724_00001838
|
| 286 |
+
ARKitScenes_test Validation_42899611_228387.721_00002318
|
| 287 |
+
ARKitScenes_test Validation_42899611_228400.716_00003098
|
| 288 |
+
ARKitScenes_test Validation_42899611_228401.716_00003158
|
| 289 |
+
ARKitScenes_test Validation_42899611_228411.712_00003758
|
| 290 |
+
ARKitScenes_test Validation_42899612_228429.521_00000229
|
| 291 |
+
ARKitScenes_test Validation_42899612_228433.520_00000469
|
| 292 |
+
ARKitScenes_test Validation_42899612_228469.522_00002630
|
| 293 |
+
ARKitScenes_test Validation_42899612_228476.519_00003050
|
| 294 |
+
ARKitScenes_test Validation_42899612_228479.518_00003230
|
| 295 |
+
ARKitScenes_test Validation_42899612_228481.517_00003350
|
| 296 |
+
ARKitScenes_test Validation_42899612_228503.525_00004671
|
| 297 |
+
ARKitScenes_test Validation_42899612_228515.521_00005391
|
| 298 |
+
ARKitScenes_test Validation_42899613_228274.016_00000102
|
| 299 |
+
ARKitScenes_test Validation_42899613_228311.018_00002323
|
| 300 |
+
ARKitScenes_test Validation_42899613_228315.017_00002563
|
| 301 |
+
ARKitScenes_test Validation_42899619_228890.872_00000157
|
| 302 |
+
ARKitScenes_test Validation_42899619_228899.869_00000697
|
| 303 |
+
ARKitScenes_test Validation_42899620_228848.672_00000097
|
| 304 |
+
ARKitScenes_test Validation_44358442_42896.098_00000343
|
| 305 |
+
ARKitScenes_test Validation_44358442_42911.092_00001243
|
| 306 |
+
ARKitScenes_test Validation_44358446_43064.895_00003398
|
| 307 |
+
ARKitScenes_test Validation_44358446_43120.888_00006759
|
| 308 |
+
ARKitScenes_test Validation_44358448_42965.802_00000955
|
| 309 |
+
ARKitScenes_test Validation_44358448_42978.797_00001735
|
| 310 |
+
ARKitScenes_test Validation_44358448_42996.790_00002815
|
| 311 |
+
ARKitScenes_test Validation_44358448_42998.789_00002935
|
| 312 |
+
ARKitScenes_test Validation_44358448_43000.805_00003056
|
| 313 |
+
ARKitScenes_test Validation_44358451_43608.890_00000217
|
| 314 |
+
ARKitScenes_test Validation_44358451_43610.889_00000337
|
| 315 |
+
ARKitScenes_test Validation_44358451_43614.888_00000577
|
| 316 |
+
ARKitScenes_test Validation_44358451_43619.886_00000877
|
| 317 |
+
ARKitScenes_test Validation_44358451_43640.894_00002138
|
| 318 |
+
ARKitScenes_test Validation_44358451_43669.881_00003878
|
| 319 |
+
ARKitScenes_test Validation_44358451_43714.896_00006580
|
| 320 |
+
ARKitScenes_test Validation_44358452_43508.598_00000517
|
| 321 |
+
ARKitScenes_test Validation_44358452_43509.598_00000577
|
| 322 |
+
ARKitScenes_test Validation_44358452_43586.583_00005198
|
| 323 |
+
ARKitScenes_test Validation_44358452_43590.598_00005439
|
| 324 |
+
ARKitScenes_test Validation_45260854_575.085_00001129
|
| 325 |
+
ARKitScenes_test Validation_45260854_578.084_00001309
|
| 326 |
+
ARKitScenes_test Validation_45260854_593.078_00002209
|
| 327 |
+
ARKitScenes_test Validation_45260854_604.090_00002870
|
| 328 |
+
ARKitScenes_test Validation_45260856_632.278_00000396
|
| 329 |
+
ARKitScenes_test Validation_45260856_635.277_00000576
|
| 330 |
+
ARKitScenes_test Validation_45260856_656.285_00001837
|
| 331 |
+
ARKitScenes_test Validation_45260856_694.286_00004118
|
| 332 |
+
ARKitScenes_test Validation_45260856_707.281_00004898
|
| 333 |
+
ARKitScenes_test Validation_45260856_713.278_00005258
|
| 334 |
+
ARKitScenes_test Validation_45260857_501.682_00000277
|
| 335 |
+
ARKitScenes_test Validation_45260857_509.679_00000757
|
| 336 |
+
ARKitScenes_test Validation_45260857_515.693_00001118
|
| 337 |
+
ARKitScenes_test Validation_45260857_518.692_00001298
|
| 338 |
+
ARKitScenes_test Validation_45260857_533.686_00002198
|
| 339 |
+
ARKitScenes_test Validation_45260857_542.682_00002738
|
| 340 |
+
ARKitScenes_test Validation_45260898_786.983_00001429
|
| 341 |
+
ARKitScenes_test Validation_45260899_887.793_00002384
|
| 342 |
+
ARKitScenes_test Validation_45260899_896.789_00002924
|
| 343 |
+
ARKitScenes_test Validation_45260900_833.481_00001813
|
| 344 |
+
ARKitScenes_test Validation_45260920_2470.417_00000228
|
| 345 |
+
ARKitScenes_test Validation_45260920_2475.432_00000529
|
| 346 |
+
ARKitScenes_test Validation_45260920_2493.425_00001609
|
| 347 |
+
ARKitScenes_test Validation_45260920_2508.419_00002509
|
| 348 |
+
ARKitScenes_test Validation_45260920_2534.425_00004070
|
| 349 |
+
ARKitScenes_test Validation_45260925_2396.830_00000157
|
| 350 |
+
ARKitScenes_test Validation_45260925_2429.817_00002137
|
| 351 |
+
ARKitScenes_test Validation_45260925_2432.816_00002317
|
| 352 |
+
ARKitScenes_test Validation_45260925_2434.831_00002438
|
| 353 |
+
ARKitScenes_test Validation_45260928_2379.220_00002498
|
| 354 |
+
ARKitScenes_test Validation_45261121_3431.070_00001034
|
| 355 |
+
ARKitScenes_test Validation_45261121_3433.069_00001154
|
| 356 |
+
ARKitScenes_test Validation_45261121_3445.064_00001874
|
| 357 |
+
ARKitScenes_test Validation_45261121_3446.063_00001934
|
| 358 |
+
ARKitScenes_test Validation_45261121_3449.062_00002114
|
| 359 |
+
ARKitScenes_test Validation_45261121_3451.061_00002234
|
| 360 |
+
ARKitScenes_test Validation_45261121_3455.060_00002474
|
| 361 |
+
ARKitScenes_test Validation_45261121_3460.058_00002774
|
| 362 |
+
ARKitScenes_test Validation_45261128_3497.159_00000457
|
| 363 |
+
ARKitScenes_test Validation_45261128_3500.158_00000637
|
| 364 |
+
ARKitScenes_test Validation_45261128_3502.157_00000757
|
| 365 |
+
ARKitScenes_test Validation_45261128_3534.160_00002678
|
| 366 |
+
ARKitScenes_test Validation_45261133_4053.263_00002797
|
| 367 |
+
ARKitScenes_test Validation_45261133_4093.263_00005198
|
| 368 |
+
ARKitScenes_test Validation_45261133_4098.261_00005498
|
| 369 |
+
ARKitScenes_test Validation_45261133_4108.273_00006099
|
| 370 |
+
ARKitScenes_test Validation_45261133_4134.262_00007659
|
| 371 |
+
ARKitScenes_test Validation_45261133_4135.262_00007719
|
| 372 |
+
ARKitScenes_test Validation_45261133_4139.260_00007959
|
| 373 |
+
ARKitScenes_test Validation_45261133_4141.259_00008079
|
| 374 |
+
ARKitScenes_test Validation_45261140_3956.270_00000583
|
| 375 |
+
ARKitScenes_test Validation_45261140_3958.269_00000703
|
| 376 |
+
ARKitScenes_test Validation_45261140_3977.261_00001843
|
| 377 |
+
ARKitScenes_test Validation_45261140_3985.274_00002324
|
| 378 |
+
ARKitScenes_test Validation_45261142_3924.466_00002335
|
| 379 |
+
ARKitScenes_test Validation_45261143_4589.702_00000643
|
| 380 |
+
ARKitScenes_test Validation_45261143_4602.714_00001424
|
| 381 |
+
ARKitScenes_test Validation_45261143_4630.702_00003104
|
| 382 |
+
ARKitScenes_test Validation_45261144_4656.008_00000337
|
| 383 |
+
ARKitScenes_test Validation_45261150_4553.301_00001657
|
| 384 |
+
ARKitScenes_test Validation_45261182_2970.729_00003398
|
| 385 |
+
ARKitScenes_test Validation_45261182_2972.728_00003518
|
| 386 |
+
ARKitScenes_test Validation_45261185_3551.484_00000283
|
| 387 |
+
ARKitScenes_test Validation_45261185_3561.480_00000883
|
| 388 |
+
ARKitScenes_test Validation_45261185_3581.489_00002084
|
| 389 |
+
ARKitScenes_test Validation_45261190_3699.892_00000823
|
| 390 |
+
ARKitScenes_test Validation_45261190_3709.888_00001423
|
| 391 |
+
ARKitScenes_test Validation_45261190_3735.895_00002984
|
| 392 |
+
ARKitScenes_test Validation_45261193_3621.190_00000169
|
| 393 |
+
ARKitScenes_test Validation_45261193_3646.180_00001669
|
| 394 |
+
ARKitScenes_test Validation_45261193_3654.194_00002150
|
| 395 |
+
ARKitScenes_test Validation_45261575_77494.516_00001538
|
| 396 |
+
ARKitScenes_test Validation_45261581_77517.507_00000097
|
| 397 |
+
ARKitScenes_test Validation_45261581_77537.515_00001298
|
| 398 |
+
ARKitScenes_test Validation_45261581_77538.515_00001358
|
| 399 |
+
ARKitScenes_test Validation_45261581_77539.514_00001418
|
| 400 |
+
ARKitScenes_test Validation_45261581_77560.505_00002678
|
| 401 |
+
ARKitScenes_test Validation_45261581_77587.511_00004299
|
| 402 |
+
ARKitScenes_test Validation_45261581_77590.510_00004479
|
| 403 |
+
ARKitScenes_test Validation_45261588_78074.609_00000637
|
| 404 |
+
ARKitScenes_test Validation_45261588_78081.606_00001057
|
| 405 |
+
ARKitScenes_test Validation_45261588_78083.605_00001177
|
| 406 |
+
ARKitScenes_test Validation_45261588_78091.602_00001657
|
| 407 |
+
ARKitScenes_test Validation_45261594_78053.201_00001477
|
| 408 |
+
ARKitScenes_test Validation_45261619_55117.682_00003176
|
| 409 |
+
ARKitScenes_test Validation_45261620_54953.280_00001297
|
| 410 |
+
ARKitScenes_test Validation_45261631_56074.403_00002858
|
| 411 |
+
ARKitScenes_test Validation_45261631_56077.402_00003038
|
| 412 |
+
ARKitScenes_test Validation_45261632_55901.188_00000678
|
| 413 |
+
ARKitScenes_test Validation_45261637_55970.494_00000877
|
| 414 |
+
ARKitScenes_test Validation_45662921_21777.915_00001627
|
| 415 |
+
ARKitScenes_test Validation_45662921_21785.928_00002108
|
| 416 |
+
ARKitScenes_test Validation_45662921_21792.926_00002528
|
| 417 |
+
ARKitScenes_test Validation_45662921_21795.924_00002708
|
| 418 |
+
ARKitScenes_test Validation_45662921_21830.926_00004809
|
| 419 |
+
ARKitScenes_test Validation_45662921_21843.921_00005589
|
| 420 |
+
ARKitScenes_test Validation_45662921_21844.921_00005649
|
| 421 |
+
ARKitScenes_test Validation_45662921_21852.917_00006129
|
| 422 |
+
ARKitScenes_test Validation_45662924_21565.920_00000637
|
| 423 |
+
ARKitScenes_test Validation_45662924_21575.916_00001237
|
| 424 |
+
ARKitScenes_test Validation_45662924_21603.921_00002918
|
| 425 |
+
ARKitScenes_test Validation_45662926_21672.326_00002048
|
| 426 |
+
ARKitScenes_test Validation_45662926_21678.323_00002408
|
| 427 |
+
ARKitScenes_test Validation_45662926_21679.323_00002468
|
| 428 |
+
ARKitScenes_test Validation_45662926_21701.330_00003789
|
| 429 |
+
ARKitScenes_test Validation_45662926_21718.323_00004809
|
| 430 |
+
ARKitScenes_test Validation_45662926_21719.323_00004869
|
| 431 |
+
ARKitScenes_test Validation_45662926_21720.322_00004929
|
| 432 |
+
ARKitScenes_test Validation_45662942_22882.675_00000037
|
| 433 |
+
ARKitScenes_test Validation_45662942_22899.668_00001057
|
| 434 |
+
ARKitScenes_test Validation_45662942_22915.661_00002017
|
| 435 |
+
ARKitScenes_test Validation_45662942_22930.672_00002918
|
| 436 |
+
ARKitScenes_test Validation_45662943_23021.267_00000157
|
| 437 |
+
ARKitScenes_test Validation_45662943_23140.268_00007300
|
| 438 |
+
ARKitScenes_test Validation_45662943_23172.254_00009220
|
| 439 |
+
ARKitScenes_test Validation_45662943_23198.260_00010781
|
| 440 |
+
ARKitScenes_test Validation_45662943_23200.259_00010901
|
| 441 |
+
ARKitScenes_test Validation_45662943_23203.258_00011081
|
| 442 |
+
ARKitScenes_test Validation_45662944_22955.761_00000817
|
| 443 |
+
ARKitScenes_test Validation_45662970_26110.534_00000247
|
| 444 |
+
ARKitScenes_test Validation_45662970_26162.530_00003368
|
| 445 |
+
ARKitScenes_test Validation_45662970_26164.529_00003488
|
| 446 |
+
ARKitScenes_test Validation_45662970_26169.527_00003788
|
| 447 |
+
ARKitScenes_test Validation_45662970_26170.527_00003848
|
| 448 |
+
ARKitScenes_test Validation_45662975_26067.134_00000163
|
| 449 |
+
ARKitScenes_test Validation_45662975_26074.131_00000583
|
| 450 |
+
ARKitScenes_test Validation_45662981_27941.098_00000289
|
| 451 |
+
ARKitScenes_test Validation_45662981_28034.095_00005871
|
| 452 |
+
ARKitScenes_test Validation_45662983_27797.888_00000097
|
| 453 |
+
ARKitScenes_test Validation_45662983_27799.887_00000217
|
| 454 |
+
ARKitScenes_test Validation_45662983_27803.886_00000457
|
| 455 |
+
ARKitScenes_test Validation_45662987_27872.192_00000349
|
| 456 |
+
ARKitScenes_test Validation_45663113_55333.668_00001778
|
| 457 |
+
ARKitScenes_test Validation_45663114_55382.864_00000397
|
| 458 |
+
ARKitScenes_test Validation_45663114_55385.863_00000577
|
| 459 |
+
ARKitScenes_test Validation_45663114_55413.868_00002258
|
| 460 |
+
ARKitScenes_test Validation_45663114_55415.867_00002378
|
| 461 |
+
ARKitScenes_test Validation_45663115_55471.860_00002558
|
| 462 |
+
ARKitScenes_test Validation_45663149_63308.707_00000757
|
| 463 |
+
ARKitScenes_test Validation_45663149_63312.706_00000997
|
| 464 |
+
ARKitScenes_test Validation_45663150_63508.953_00000108
|
| 465 |
+
ARKitScenes_test Validation_45663150_63515.950_00000528
|
| 466 |
+
ARKitScenes_test Validation_45663150_63519.949_00000768
|
| 467 |
+
ARKitScenes_test Validation_45663150_63548.954_00002509
|
| 468 |
+
ARKitScenes_test Validation_45663150_63575.960_00004130
|
| 469 |
+
ARKitScenes_test Validation_45663150_63579.958_00004370
|
| 470 |
+
ARKitScenes_test Validation_45663150_63582.957_00004550
|
| 471 |
+
ARKitScenes_test Validation_45663154_63400.863_00000133
|
| 472 |
+
ARKitScenes_test Validation_45663154_63407.860_00000553
|
| 473 |
+
ARKitScenes_test Validation_45663154_63408.859_00000613
|
| 474 |
+
ARKitScenes_test Validation_45663154_63435.849_00002233
|
| 475 |
+
ARKitScenes_test Validation_45663154_63436.848_00002293
|
| 476 |
+
ARKitScenes_test Validation_45663154_63451.859_00003194
|
| 477 |
+
ARKitScenes_test Validation_45663164_65934.817_00000234
|
| 478 |
+
ARKitScenes_test Validation_45663164_65937.816_00000414
|
| 479 |
+
ARKitScenes_test Validation_45663164_65954.825_00001435
|
| 480 |
+
ARKitScenes_test Validation_45663164_65965.821_00002095
|
| 481 |
+
ARKitScenes_test Validation_45663164_65971.819_00002455
|
| 482 |
+
ARKitScenes_test Validation_45663164_65981.815_00003055
|
| 483 |
+
ARKitScenes_test Validation_45663164_65990.828_00003596
|
| 484 |
+
ARKitScenes_test Validation_45663164_66002.823_00004316
|
| 485 |
+
ARKitScenes_test Validation_45663165_65825.027_00001238
|
| 486 |
+
ARKitScenes_test Validation_45663165_65826.026_00001298
|
| 487 |
+
ARKitScenes_test Validation_45663175_65888.618_00001033
|
| 488 |
+
ARKitScenes_test Validation_45663175_65895.616_00001453
|
| 489 |
+
ARKitScenes_test Validation_45663175_65900.614_00001753
|
| 490 |
+
ARKitScenes_test Validation_47115452_13059.567_00000217
|
| 491 |
+
ARKitScenes_test Validation_47115452_13068.580_00000758
|
| 492 |
+
ARKitScenes_test Validation_47115452_13087.572_00001898
|
| 493 |
+
ARKitScenes_test Validation_47115452_13111.579_00003339
|
| 494 |
+
ARKitScenes_test Validation_47115452_13118.576_00003759
|
| 495 |
+
ARKitScenes_test Validation_47115452_13119.575_00003819
|
| 496 |
+
ARKitScenes_test Validation_47115460_13170.771_00002029
|
| 497 |
+
ARKitScenes_test Validation_47115463_13276.477_00003458
|
| 498 |
+
ARKitScenes_test Validation_47115463_13302.466_00005018
|
| 499 |
+
ARKitScenes_test Validation_47115469_14015.292_00000277
|
| 500 |
+
ARKitScenes_test Validation_47115469_14028.286_00001057
|
| 501 |
+
ARKitScenes_test Validation_47115469_14030.285_00001177
|
| 502 |
+
ARKitScenes_test Validation_47115473_14110.785_00002330
|
| 503 |
+
ARKitScenes_test Validation_47115473_14130.777_00003530
|
| 504 |
+
ARKitScenes_test Validation_47115473_14143.788_00004311
|
| 505 |
+
ARKitScenes_test Validation_47115473_14173.776_00006111
|
| 506 |
+
ARKitScenes_test Validation_47115474_13964.679_00000282
|
| 507 |
+
ARKitScenes_test Validation_47115474_13981.689_00001303
|
| 508 |
+
ARKitScenes_test Validation_47204552_3362.090_00000577
|
| 509 |
+
ARKitScenes_test Validation_47204552_3393.094_00002438
|
| 510 |
+
ARKitScenes_test Validation_47204552_3399.091_00002798
|
| 511 |
+
ARKitScenes_test Validation_47204552_3401.091_00002918
|
| 512 |
+
ARKitScenes_test Validation_47204552_3402.090_00002978
|
| 513 |
+
ARKitScenes_test Validation_47204552_3415.101_00003759
|
| 514 |
+
ARKitScenes_test Validation_47204552_3423.098_00004239
|
| 515 |
+
ARKitScenes_test Validation_47204552_3427.097_00004479
|
| 516 |
+
ARKitScenes_test Validation_47204552_3438.092_00005139
|
| 517 |
+
ARKitScenes_test Validation_47204552_3450.087_00005859
|
| 518 |
+
ARKitScenes_test Validation_47204552_3478.092_00007540
|
| 519 |
+
ARKitScenes_test Validation_47204554_3222.198_00000697
|
| 520 |
+
ARKitScenes_test Validation_47204554_3262.198_00003098
|
| 521 |
+
ARKitScenes_test Validation_47204554_3276.192_00003938
|
| 522 |
+
ARKitScenes_test Validation_47204556_3287.388_00000096
|
| 523 |
+
ARKitScenes_test Validation_47204556_3289.387_00000216
|
| 524 |
+
ARKitScenes_test Validation_47204556_3296.401_00000637
|
| 525 |
+
ARKitScenes_test Validation_47204556_3318.392_00001957
|
| 526 |
+
ARKitScenes_test Validation_47204556_3339.400_00003218
|
| 527 |
+
ARKitScenes_test Validation_47204559_4004.548_00000997
|
| 528 |
+
ARKitScenes_test Validation_47204559_4005.547_00001057
|
| 529 |
+
ARKitScenes_test Validation_47204563_4035.451_00000157
|
| 530 |
+
ARKitScenes_test Validation_47204566_4079.050_00000217
|
| 531 |
+
ARKitScenes_test Validation_47204566_4108.038_00001957
|
| 532 |
+
ARKitScenes_test Validation_47204566_4112.036_00002197
|
| 533 |
+
ARKitScenes_test Validation_47204566_4122.049_00002798
|
| 534 |
+
ARKitScenes_test Validation_47204566_4127.047_00003098
|
| 535 |
+
ARKitScenes_test Validation_47204573_4957.623_00000038
|
| 536 |
+
ARKitScenes_test Validation_47204573_4997.607_00002438
|
| 537 |
+
ARKitScenes_test Validation_47204573_5033.609_00004599
|
| 538 |
+
ARKitScenes_test Validation_47204573_5037.607_00004839
|
| 539 |
+
ARKitScenes_test Validation_47204575_4843.221_00000529
|
| 540 |
+
ARKitScenes_test Validation_47204575_4872.209_00002269
|
| 541 |
+
ARKitScenes_test Validation_47204575_4874.208_00002389
|
| 542 |
+
ARKitScenes_test Validation_47204578_4911.209_00001063
|
| 543 |
+
ARKitScenes_test Validation_47204578_4929.219_00002144
|
| 544 |
+
ARKitScenes_test Validation_47204605_2404.299_00000121
|
| 545 |
+
ARKitScenes_test Validation_47204605_2412.295_00000601
|
| 546 |
+
ARKitScenes_test Validation_47204605_2419.293_00001021
|
| 547 |
+
ARKitScenes_test Validation_47204605_2420.292_00001081
|
| 548 |
+
ARKitScenes_test Validation_47204605_2431.305_00001742
|
| 549 |
+
ARKitScenes_test Validation_47204605_2436.303_00002042
|
| 550 |
+
ARKitScenes_test Validation_47204605_2440.301_00002282
|
| 551 |
+
ARKitScenes_test Validation_47204607_2528.799_00000517
|
| 552 |
+
ARKitScenes_test Validation_47204607_2560.803_00002438
|
| 553 |
+
ARKitScenes_test Validation_47204607_2561.803_00002498
|
| 554 |
+
ARKitScenes_test Validation_47204607_2569.800_00002978
|
| 555 |
+
ARKitScenes_test Validation_47204607_2573.798_00003218
|
| 556 |
+
ARKitScenes_test Validation_47204607_2574.798_00003278
|
| 557 |
+
ARKitScenes_test Validation_47204607_2577.797_00003458
|
| 558 |
+
ARKitScenes_test Validation_47204607_2583.794_00003818
|
| 559 |
+
ARKitScenes_test Validation_47204607_2591.791_00004298
|
| 560 |
+
ARKitScenes_test Validation_47204607_2626.794_00006399
|
| 561 |
+
ARKitScenes_test Validation_47204607_2634.791_00006879
|
| 562 |
+
ARKitScenes_test Validation_47204607_2664.796_00008680
|
| 563 |
+
ARKitScenes_test Validation_47204609_2453.013_00000038
|
| 564 |
+
ARKitScenes_test Validation_47204609_2485.000_00001958
|
| 565 |
+
ARKitScenes_test Validation_47331068_131078.065_00000817
|
| 566 |
+
ARKitScenes_test Validation_47331069_131220.357_00004598
|
| 567 |
+
ARKitScenes_test Validation_47331071_131007.360_00000817
|
| 568 |
+
ARKitScenes_test Validation_47331265_90013.317_00000134
|
| 569 |
+
ARKitScenes_test Validation_47331265_90032.293_00001273
|
| 570 |
+
ARKitScenes_test Validation_47331265_90067.295_00003374
|
| 571 |
+
ARKitScenes_test Validation_47331266_89823.527_00000141
|
| 572 |
+
ARKitScenes_test Validation_47331266_89825.493_00000259
|
| 573 |
+
ARKitScenes_test Validation_47331266_89873.490_00003140
|
| 574 |
+
ARKitScenes_test Validation_47331319_179040.028_00000234
|
| 575 |
+
ARKitScenes_test Validation_47331319_179072.032_00002155
|
| 576 |
+
ARKitScenes_test Validation_47331322_179115.348_00000110
|
| 577 |
+
ARKitScenes_test Validation_47331322_179135.339_00001310
|
| 578 |
+
ARKitScenes_test Validation_47331324_179185.452_00000122
|
| 579 |
+
ARKitScenes_test Validation_47331324_179188.434_00000301
|
| 580 |
+
ARKitScenes_test Validation_47331324_179200.430_00001021
|
| 581 |
+
ARKitScenes_test Validation_47331324_179214.424_00001861
|
| 582 |
+
ARKitScenes_test Validation_47331324_179216.440_00001982
|
| 583 |
+
ARKitScenes_test Validation_47331651_22862.230_00000757
|
| 584 |
+
ARKitScenes_test Validation_47331651_22892.235_00002558
|
| 585 |
+
ARKitScenes_test Validation_47331651_22893.235_00002618
|
| 586 |
+
ARKitScenes_test Validation_47331651_22920.224_00004238
|
| 587 |
+
ARKitScenes_test Validation_47331651_22973.235_00007420
|
| 588 |
+
ARKitScenes_test Validation_47331653_23010.337_00001538
|
| 589 |
+
ARKitScenes_test Validation_47331653_23016.334_00001898
|
| 590 |
+
ARKitScenes_test Validation_47331653_23030.329_00002738
|
| 591 |
+
ARKitScenes_test Validation_47331653_23034.327_00002978
|
| 592 |
+
ARKitScenes_test Validation_47331653_23076.326_00005499
|
| 593 |
+
ARKitScenes_test Validation_47331653_23080.325_00005739
|
| 594 |
+
ARKitScenes_test Validation_47331653_23090.337_00006340
|
| 595 |
+
ARKitScenes_test Validation_47331653_23092.337_00006460
|
| 596 |
+
ARKitScenes_test Validation_47331653_23132.320_00008860
|
| 597 |
+
ARKitScenes_test Validation_47331654_22744.828_00000517
|
| 598 |
+
ARKitScenes_test Validation_47331654_22746.827_00000637
|
| 599 |
+
ARKitScenes_test Validation_47331654_22751.825_00000937
|
| 600 |
+
ARKitScenes_test Validation_47331654_22772.834_00002198
|
| 601 |
+
ARKitScenes_test Validation_47331654_22781.830_00002738
|
| 602 |
+
ARKitScenes_test Validation_47331654_22787.827_00003098
|
| 603 |
+
ARKitScenes_test Validation_47331654_22797.823_00003698
|
| 604 |
+
ARKitScenes_test Validation_47331654_22833.825_00005859
|
| 605 |
+
ARKitScenes_test Validation_47331661_23629.582_00002630
|
| 606 |
+
ARKitScenes_test Validation_47331661_23690.590_00006292
|
| 607 |
+
ARKitScenes_test Validation_47331662_23429.080_00000516
|
| 608 |
+
ARKitScenes_test Validation_47331662_23466.082_00002737
|
| 609 |
+
ARKitScenes_test Validation_47331662_23470.080_00002977
|
| 610 |
+
ARKitScenes_test Validation_47331662_23472.079_00003097
|
| 611 |
+
ARKitScenes_test Validation_47331988_170658.548_00003824
|
| 612 |
+
ARKitScenes_test Validation_47331988_170671.559_00004605
|
| 613 |
+
ARKitScenes_test Validation_47331988_170673.559_00004725
|
| 614 |
+
ARKitScenes_test Validation_47331988_170674.558_00004785
|
| 615 |
+
ARKitScenes_test Validation_47331988_170681.555_00005205
|
| 616 |
+
ARKitScenes_test Validation_47331989_170440.853_00001597
|
| 617 |
+
ARKitScenes_test Validation_47331990_170486.951_00000408
|
| 618 |
+
ARKitScenes_test Validation_47331990_170529.950_00002989
|
| 619 |
+
ARKitScenes_test Validation_47331990_170567.951_00005270
|
| 620 |
+
ARKitScenes_test Validation_47331990_170578.963_00005931
|
| 621 |
+
ARKitScenes_test Validation_47332004_171938.432_00000997
|
| 622 |
+
ARKitScenes_test Validation_47332005_172095.735_00001963
|
| 623 |
+
ARKitScenes_test Validation_47332005_172102.732_00002383
|
| 624 |
+
ARKitScenes_test Validation_47332005_172149.729_00005204
|
| 625 |
+
ARKitScenes_test Validation_47332885_7500.859_00001003
|
| 626 |
+
ARKitScenes_test Validation_47332885_7517.869_00002024
|
| 627 |
+
ARKitScenes_test Validation_47332886_7549.256_00000474
|
| 628 |
+
ARKitScenes_test Validation_47332886_7550.256_00000534
|
| 629 |
+
ARKitScenes_test Validation_47332886_7562.267_00001255
|
| 630 |
+
ARKitScenes_test Validation_47332886_7567.265_00001555
|
| 631 |
+
ARKitScenes_test Validation_47332886_7582.259_00002455
|
| 632 |
+
ARKitScenes_test Validation_47332890_7653.364_00003398
|
| 633 |
+
ARKitScenes_test Validation_47332890_7662.360_00003938
|
| 634 |
+
ARKitScenes_test Validation_47332890_7663.360_00003998
|
| 635 |
+
ARKitScenes_test Validation_47332890_7667.358_00004238
|
| 636 |
+
ARKitScenes_test Validation_47332890_7668.358_00004298
|
| 637 |
+
ARKitScenes_test Validation_47332893_9315.687_00007719
|
| 638 |
+
ARKitScenes_test Validation_47332895_9057.792_00000715
|
| 639 |
+
ARKitScenes_test Validation_47332895_9066.788_00001255
|
| 640 |
+
ARKitScenes_test Validation_47332899_9174.194_00003308
|
| 641 |
+
ARKitScenes_test Validation_47332899_9175.194_00003368
|
| 642 |
+
ARKitScenes_test Validation_47332899_9177.193_00003488
|
| 643 |
+
ARKitScenes_test Validation_47332899_9179.192_00003608
|
| 644 |
+
ARKitScenes_test Validation_47332901_11032.594_00000985
|
| 645 |
+
ARKitScenes_test Validation_47332901_11062.598_00002786
|
| 646 |
+
ARKitScenes_test Validation_47332904_11093.386_00000943
|
| 647 |
+
ARKitScenes_test Validation_47332904_11096.385_00001123
|
| 648 |
+
ARKitScenes_test Validation_47332904_11149.396_00004305
|
| 649 |
+
ARKitScenes_test Validation_47332905_10961.889_00000709
|
| 650 |
+
ARKitScenes_test Validation_47332905_10964.888_00000889
|
| 651 |
+
ARKitScenes_test Validation_47332910_12198.325_00000708
|
| 652 |
+
ARKitScenes_test Validation_47332911_12283.823_00001483
|
| 653 |
+
ARKitScenes_test Validation_47332911_12347.831_00005325
|
| 654 |
+
ARKitScenes_test Validation_47332911_12351.829_00005565
|
| 655 |
+
ARKitScenes_test Validation_47332915_13476.969_00000643
|
| 656 |
+
ARKitScenes_test Validation_47332915_13505.974_00002384
|
| 657 |
+
ARKitScenes_test Validation_47332915_13538.977_00004365
|
| 658 |
+
ARKitScenes_test Validation_47332915_13547.973_00004905
|
| 659 |
+
ARKitScenes_test Validation_47332915_13552.971_00005205
|
| 660 |
+
ARKitScenes_test Validation_47332915_13560.968_00005685
|
| 661 |
+
ARKitScenes_test Validation_47332918_13321.866_00000349
|
| 662 |
+
ARKitScenes_test Validation_47332918_13363.865_00002870
|
| 663 |
+
ARKitScenes_test Validation_47333440_53806.225_00000757
|
| 664 |
+
ARKitScenes_test Validation_47333443_53755.129_00000397
|
| 665 |
+
ARKitScenes_test Validation_47333443_53760.127_00000697
|
| 666 |
+
ARKitScenes_test Validation_47333452_54966.217_00000883
|
| 667 |
+
ARKitScenes_test Validation_47333452_54993.223_00002504
|
| 668 |
+
ARKitScenes_test Validation_47333452_54995.222_00002624
|
| 669 |
+
ARKitScenes_test Validation_47333456_55028.108_00001237
|
| 670 |
+
ARKitScenes_test Validation_47333456_55030.107_00001357
|
| 671 |
+
ARKitScenes_test Validation_47333456_55035.122_00001658
|
| 672 |
+
ARKitScenes_test Validation_47333456_55058.112_00003038
|
| 673 |
+
ARKitScenes_test Validation_47333456_55067.109_00003578
|
| 674 |
+
ARKitScenes_test Validation_47333898_81325.974_00000397
|
| 675 |
+
ARKitScenes_test Validation_47333898_81346.966_00001657
|
| 676 |
+
ARKitScenes_test Validation_47333898_81372.972_00003218
|
| 677 |
+
ARKitScenes_test Validation_47333898_81377.970_00003518
|
| 678 |
+
ARKitScenes_test Validation_47333898_81378.969_00003578
|
| 679 |
+
ARKitScenes_test Validation_47333899_81255.070_00000877
|
| 680 |
+
ARKitScenes_test Validation_47333899_81271.063_00001837
|
| 681 |
+
ARKitScenes_test Validation_47333899_81306.066_00003938
|
| 682 |
+
ARKitScenes_test Validation_47333904_81416.371_00000877
|
| 683 |
+
ARKitScenes_test Validation_47333904_81425.367_00001417
|
| 684 |
+
ARKitScenes_test Validation_47333904_81432.364_00001837
|
| 685 |
+
ARKitScenes_test Validation_47333916_82430.565_00003759
|
| 686 |
+
ARKitScenes_test Validation_47333916_82435.563_00004059
|
| 687 |
+
ARKitScenes_test Validation_47333918_82512.465_00003404
|
| 688 |
+
ARKitScenes_test Validation_47333918_82533.457_00004664
|
| 689 |
+
ARKitScenes_test Validation_47333920_82642.262_00005856
|
| 690 |
+
ARKitScenes_test Validation_47333920_82666.253_00007296
|
| 691 |
+
ARKitScenes_test Validation_47333923_83604.501_00000396
|
| 692 |
+
ARKitScenes_test Validation_47333925_83573.613_00001964
|
| 693 |
+
ARKitScenes_test Validation_47333927_84005.212_00000944
|
| 694 |
+
ARKitScenes_test Validation_47333927_84013.209_00001424
|
| 695 |
+
ARKitScenes_test Validation_47333927_84020.206_00001844
|
| 696 |
+
ARKitScenes_test Validation_47333927_84025.204_00002144
|
| 697 |
+
ARKitScenes_test Validation_47333927_84028.203_00002324
|
| 698 |
+
ARKitScenes_test Validation_47333931_83940.705_00001658
|
| 699 |
+
ARKitScenes_test Validation_47333931_83943.704_00001838
|
| 700 |
+
ARKitScenes_test Validation_47333932_84109.203_00002798
|
| 701 |
+
ARKitScenes_test Validation_47333932_84184.206_00007300
|
| 702 |
+
ARKitScenes_test Validation_47333934_85556.772_00000469
|
| 703 |
+
ARKitScenes_test Validation_47333934_85572.765_00001429
|
| 704 |
+
ARKitScenes_test Validation_47333934_85623.778_00004491
|
| 705 |
+
ARKitScenes_test Validation_47333934_85636.773_00005271
|
| 706 |
+
ARKitScenes_test Validation_47333934_85638.772_00005391
|
| 707 |
+
ARKitScenes_test Validation_47333934_85646.769_00005871
|
| 708 |
+
ARKitScenes_test Validation_47333934_85659.763_00006651
|
| 709 |
+
ARKitScenes_test Validation_47333937_85498.362_00000540
|
| 710 |
+
ARKitScenes_test Validation_47333937_85501.378_00000721
|
| 711 |
+
ARKitScenes_test Validation_47333937_85528.367_00002341
|
| 712 |
+
ARKitScenes_test Validation_47333937_85541.378_00003122
|
| 713 |
+
ARKitScenes_test Validation_47333937_85543.377_00003242
|
| 714 |
+
ARKitScenes_test Validation_47333940_85418.978_00000517
|
| 715 |
+
ARKitScenes_test Validation_47333940_85420.977_00000637
|
| 716 |
+
ARKitScenes_test Validation_47333940_85422.976_00000757
|
| 717 |
+
ARKitScenes_test Validation_47333940_85438.969_00001717
|
| 718 |
+
ARKitScenes_test Validation_47333940_85480.969_00004238
|
| 719 |
+
ARKitScenes_test Validation_47334105_92943.745_00003650
|
| 720 |
+
ARKitScenes_test Validation_47334105_92981.746_00005931
|
| 721 |
+
ARKitScenes_test Validation_47334105_93005.753_00007372
|
| 722 |
+
ARKitScenes_test Validation_47334239_111152.021_00002918
|
| 723 |
+
ARKitScenes_test Validation_47334240_111223.425_00003818
|
| 724 |
+
ARKitScenes_test Validation_47429912_136038.580_00000937
|
| 725 |
+
ARKitScenes_test Validation_47429912_136057.589_00002078
|
| 726 |
+
ARKitScenes_test Validation_47429922_136625.256_00000577
|
| 727 |
+
ARKitScenes_test Validation_47429971_16341.763_00003758
|
| 728 |
+
ARKitScenes_test Validation_47429995_16689.687_00000229
|
| 729 |
+
ARKitScenes_test Validation_47430024_18796.033_00001657
|
| 730 |
+
ARKitScenes_test Validation_47430026_18833.034_00001297
|
| 731 |
+
ARKitScenes_test Validation_47430026_18844.046_00001958
|
| 732 |
+
ARKitScenes_test Validation_47430033_19430.380_00001363
|
| 733 |
+
ARKitScenes_test Validation_47430036_19503.284_00000277
|
| 734 |
+
ARKitScenes_test Validation_47430043_20580.283_00000156
|
| 735 |
+
ARKitScenes_test Validation_47430045_20645.389_00001177
|
| 736 |
+
ARKitScenes_test Validation_47430047_21106.832_00000937
|
| 737 |
+
ARKitScenes_test Validation_47430047_21107.832_00000997
|
| 738 |
+
ARKitScenes_test Validation_47895341_96434.329_00002918
|
| 739 |
+
ARKitScenes_test Validation_47895348_96489.523_00001837
|
| 740 |
+
ARKitScenes_test Validation_47895348_96510.515_00003097
|
| 741 |
+
ARKitScenes_test Validation_47895348_96535.521_00004598
|
| 742 |
+
ARKitScenes_test Validation_47895348_96543.518_00005078
|
| 743 |
+
ARKitScenes_test Validation_47895350_96587.817_00002197
|
| 744 |
+
ARKitScenes_test Validation_47895350_96593.814_00002557
|
| 745 |
+
ARKitScenes_test Validation_47895350_96598.812_00002857
|
| 746 |
+
ARKitScenes_test Validation_47895350_96692.824_00008500
|
| 747 |
+
ARKitScenes_test Validation_47895355_98785.640_00001898
|
| 748 |
+
ARKitScenes_test Validation_47895355_98786.640_00001958
|
| 749 |
+
ARKitScenes_test Validation_47895355_98811.630_00003458
|
| 750 |
+
ARKitScenes_test Validation_47895355_98813.629_00003578
|
| 751 |
+
ARKitScenes_test Validation_47895355_98817.627_00003818
|
| 752 |
+
ARKitScenes_test Validation_47895355_98873.638_00007180
|
| 753 |
+
ARKitScenes_test Validation_47895355_98884.633_00007840
|
| 754 |
+
ARKitScenes_test Validation_47895355_98890.631_00008200
|
| 755 |
+
ARKitScenes_test Validation_47895355_98898.628_00008680
|
| 756 |
+
ARKitScenes_test Validation_47895355_98911.639_00009461
|
| 757 |
+
ARKitScenes_test Validation_47895355_98953.639_00011982
|
| 758 |
+
ARKitScenes_test Validation_47895364_100176.943_00000457
|
| 759 |
+
ARKitScenes_test Validation_47895364_100304.941_00008140
|
| 760 |
+
ARKitScenes_test Validation_47895365_100109.337_00000637
|
| 761 |
+
ARKitScenes_test Validation_47895556_457844.157_00003123
|
| 762 |
+
ARKitScenes_test Validation_47895783_387008.265_00000337
|
| 763 |
+
ARKitScenes_test Validation_47895783_387020.260_00001057
|
| 764 |
+
ARKitScenes_test Validation_48018367_389470.584_00001538
|
| 765 |
+
ARKitScenes_test Validation_48018368_389568.277_00002090
|
| 766 |
+
ARKitScenes_test Validation_48018368_389569.276_00002150
|
| 767 |
+
ARKitScenes_test Validation_48018368_389589.268_00003350
|
| 768 |
+
ARKitScenes_test Validation_48018372_389499.672_00000577
|
| 769 |
+
ARKitScenes_test Validation_48018379_390028.141_00001958
|
| 770 |
+
ARKitScenes_test Validation_48018559_3343.849_00000157
|
| 771 |
+
ARKitScenes_test Validation_48018560_3398.743_00000397
|
| 772 |
+
ARKitScenes_test Validation_48018560_3417.752_00001538
|
| 773 |
+
ARKitScenes_test Validation_48018566_3878.430_00001303
|
| 774 |
+
ARKitScenes_test Validation_48018571_3900.055_00000140
|
| 775 |
+
ARKitScenes_test Validation_48018572_3834.032_00002288
|
| 776 |
+
ARKitScenes_test Validation_48018730_1763.469_00005739
|
| 777 |
+
ARKitScenes_test Validation_48018732_1944.579_00002060
|
| 778 |
+
ARKitScenes_test Validation_48018732_1949.577_00002360
|
| 779 |
+
ARKitScenes_test Validation_48018732_1965.571_00003320
|
| 780 |
+
ARKitScenes_test Validation_48018732_1994.575_00005061
|
| 781 |
+
ARKitScenes_test Validation_48458484_111570.316_00000037
|
| 782 |
+
ARKitScenes_test Validation_48458484_111592.307_00001357
|
| 783 |
+
ARKitScenes_test Validation_48458484_111596.305_00001597
|
| 784 |
+
ARKitScenes_test Validation_48458650_7065.603_00000396
|
| 785 |
+
ARKitScenes_test Validation_48458650_7072.616_00000817
|
| 786 |
+
ARKitScenes_test Validation_48458650_7110.617_00003098
|
| 787 |
+
ARKitScenes_test Validation_48458656_7631.585_00000276
|
| 788 |
+
ARKitScenes_test Validation_48458656_7633.584_00000396
|
| 789 |
+
ARKitScenes_test Validation_48458656_7638.582_00000696
|
| 790 |
+
ARKitScenes_test Validation_48458656_7648.595_00001297
|
| 791 |
+
ARKitScenes_test Validation_48458656_7649.594_00001357
|
| 792 |
+
ARKitScenes_test Validation_48458657_7594.684_00000642
|
| 793 |
+
ARKitScenes_test Validation_48458657_7609.694_00001543
|
| 794 |
+
ARKitScenes_test Validation_48458660_8266.963_00001418
|
| 795 |
+
ARKitScenes_test Validation_48458667_8357.659_00000157
|
| 796 |
+
ARKitScenes_test Validation_48458667_8361.657_00000397
|
| 797 |
+
ARKitScenes_test Validation_48458667_8379.649_00001477
|
| 798 |
+
ARKitScenes_test Validation_48458667_8382.648_00001657
|
| 799 |
+
ARKitScenes_test Validation_48458667_8394.660_00002378
|
| 800 |
+
ARKitScenes_test Validation_48458667_8420.649_00003938
|
external/WildCamera/splits/arkitscenes_train.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
external/WildCamera/splits/arkitscenes_val.txt
ADDED
|
@@ -0,0 +1,800 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ARKitScenes_test Validation_41069021_370.367_00003945
|
| 2 |
+
ARKitScenes_test Validation_41069021_387.360_00004965
|
| 3 |
+
ARKitScenes_test Validation_41069021_394.358_00005385
|
| 4 |
+
ARKitScenes_test Validation_41069021_406.369_00006106
|
| 5 |
+
ARKitScenes_test Validation_41069021_409.368_00006286
|
| 6 |
+
ARKitScenes_test Validation_41069021_425.362_00007246
|
| 7 |
+
ARKitScenes_test Validation_41069021_426.361_00007306
|
| 8 |
+
ARKitScenes_test Validation_41069021_430.359_00007546
|
| 9 |
+
ARKitScenes_test Validation_41069021_431.359_00007606
|
| 10 |
+
ARKitScenes_test Validation_41069021_444.370_00008387
|
| 11 |
+
ARKitScenes_test Validation_41069021_471.359_00010007
|
| 12 |
+
ARKitScenes_test Validation_41069021_488.369_00011028
|
| 13 |
+
ARKitScenes_test Validation_41069021_492.367_00011268
|
| 14 |
+
ARKitScenes_test Validation_41069025_607.870_00001117
|
| 15 |
+
ARKitScenes_test Validation_41069025_649.869_00003638
|
| 16 |
+
ARKitScenes_test Validation_41069025_701.864_00006759
|
| 17 |
+
ARKitScenes_test Validation_41069025_711.860_00007359
|
| 18 |
+
ARKitScenes_test Validation_41069025_717.858_00007719
|
| 19 |
+
ARKitScenes_test Validation_41069025_730.869_00008500
|
| 20 |
+
ARKitScenes_test Validation_41069025_735.867_00008800
|
| 21 |
+
ARKitScenes_test Validation_41069042_3049.237_00000385
|
| 22 |
+
ARKitScenes_test Validation_41069042_3069.228_00001585
|
| 23 |
+
ARKitScenes_test Validation_41069042_3071.227_00001705
|
| 24 |
+
ARKitScenes_test Validation_41069042_3105.230_00003746
|
| 25 |
+
ARKitScenes_test Validation_41069043_2860.631_00003698
|
| 26 |
+
ARKitScenes_test Validation_41069043_2893.634_00005679
|
| 27 |
+
ARKitScenes_test Validation_41069046_2921.323_00000636
|
| 28 |
+
ARKitScenes_test Validation_41069046_2923.339_00000757
|
| 29 |
+
ARKitScenes_test Validation_41069046_2925.338_00000877
|
| 30 |
+
ARKitScenes_test Validation_41069046_2930.336_00001177
|
| 31 |
+
ARKitScenes_test Validation_41069046_2960.323_00002977
|
| 32 |
+
ARKitScenes_test Validation_41069046_2963.322_00003157
|
| 33 |
+
ARKitScenes_test Validation_41069046_2980.332_00004178
|
| 34 |
+
ARKitScenes_test Validation_41142278_4002.081_00000037
|
| 35 |
+
ARKitScenes_test Validation_41142278_4042.081_00002438
|
| 36 |
+
ARKitScenes_test Validation_41142280_3836.982_00000877
|
| 37 |
+
ARKitScenes_test Validation_42444946_223633.870_00001597
|
| 38 |
+
ARKitScenes_test Validation_42444946_223675.869_00004118
|
| 39 |
+
ARKitScenes_test Validation_42444946_223682.866_00004538
|
| 40 |
+
ARKitScenes_test Validation_42444949_223780.677_00004479
|
| 41 |
+
ARKitScenes_test Validation_42444949_223783.675_00004659
|
| 42 |
+
ARKitScenes_test Validation_42444950_223809.565_00000697
|
| 43 |
+
ARKitScenes_test Validation_42444950_223855.563_00003458
|
| 44 |
+
ARKitScenes_test Validation_42444966_40343.793_00000883
|
| 45 |
+
ARKitScenes_test Validation_42444966_40380.794_00003104
|
| 46 |
+
ARKitScenes_test Validation_42444966_40384.793_00003344
|
| 47 |
+
ARKitScenes_test Validation_42444966_40393.789_00003884
|
| 48 |
+
ARKitScenes_test Validation_42444966_40402.785_00004424
|
| 49 |
+
ARKitScenes_test Validation_42444966_40406.783_00004664
|
| 50 |
+
ARKitScenes_test Validation_42444966_40407.783_00004724
|
| 51 |
+
ARKitScenes_test Validation_42444968_40245.400_00000049
|
| 52 |
+
ARKitScenes_test Validation_42444968_40246.383_00000108
|
| 53 |
+
ARKitScenes_test Validation_42444968_40252.381_00000468
|
| 54 |
+
ARKitScenes_test Validation_42444968_40258.395_00000829
|
| 55 |
+
ARKitScenes_test Validation_42444968_40297.396_00003170
|
| 56 |
+
ARKitScenes_test Validation_42444968_40300.394_00003350
|
| 57 |
+
ARKitScenes_test Validation_42444968_40312.389_00004070
|
| 58 |
+
ARKitScenes_test Validation_42444976_40143.393_00000157
|
| 59 |
+
ARKitScenes_test Validation_42444976_40169.382_00001717
|
| 60 |
+
ARKitScenes_test Validation_42444976_40170.382_00001777
|
| 61 |
+
ARKitScenes_test Validation_42444976_40171.398_00001838
|
| 62 |
+
ARKitScenes_test Validation_42444976_40182.393_00002498
|
| 63 |
+
ARKitScenes_test Validation_42444976_40186.392_00002738
|
| 64 |
+
ARKitScenes_test Validation_42444976_40198.387_00003458
|
| 65 |
+
ARKitScenes_test Validation_42444976_40199.386_00003518
|
| 66 |
+
ARKitScenes_test Validation_42445021_48884.258_00000276
|
| 67 |
+
ARKitScenes_test Validation_42445021_48887.257_00000456
|
| 68 |
+
ARKitScenes_test Validation_42445022_49008.276_00000049
|
| 69 |
+
ARKitScenes_test Validation_42445022_49014.256_00000408
|
| 70 |
+
ARKitScenes_test Validation_42445022_49015.256_00000468
|
| 71 |
+
ARKitScenes_test Validation_42445022_49035.265_00001669
|
| 72 |
+
ARKitScenes_test Validation_42445022_49070.268_00003770
|
| 73 |
+
ARKitScenes_test Validation_42445028_49607.697_00000349
|
| 74 |
+
ARKitScenes_test Validation_42445028_49609.696_00000469
|
| 75 |
+
ARKitScenes_test Validation_42445028_49655.695_00003230
|
| 76 |
+
ARKitScenes_test Validation_42445429_50330.084_00000817
|
| 77 |
+
ARKitScenes_test Validation_42445429_50333.083_00000997
|
| 78 |
+
ARKitScenes_test Validation_42445441_50300.796_00005196
|
| 79 |
+
ARKitScenes_test Validation_42445448_50410.586_00000115
|
| 80 |
+
ARKitScenes_test Validation_42445448_50486.589_00004677
|
| 81 |
+
ARKitScenes_test Validation_42445448_50494.586_00005157
|
| 82 |
+
ARKitScenes_test Validation_42445991_75264.421_00000522
|
| 83 |
+
ARKitScenes_test Validation_42445991_75304.421_00002923
|
| 84 |
+
ARKitScenes_test Validation_42445991_75326.429_00004244
|
| 85 |
+
ARKitScenes_test Validation_42446038_347909.974_00003044
|
| 86 |
+
ARKitScenes_test Validation_42446038_347925.967_00004004
|
| 87 |
+
ARKitScenes_test Validation_42446038_347928.983_00004185
|
| 88 |
+
ARKitScenes_test Validation_42446100_77913.295_00000397
|
| 89 |
+
ARKitScenes_test Validation_42446100_77962.292_00003338
|
| 90 |
+
ARKitScenes_test Validation_42446100_77966.290_00003578
|
| 91 |
+
ARKitScenes_test Validation_42446100_77994.296_00005259
|
| 92 |
+
ARKitScenes_test Validation_42446100_77997.295_00005439
|
| 93 |
+
ARKitScenes_test Validation_42446100_77999.294_00005559
|
| 94 |
+
ARKitScenes_test Validation_42446100_78006.291_00005979
|
| 95 |
+
ARKitScenes_test Validation_42446100_78007.291_00006039
|
| 96 |
+
ARKitScenes_test Validation_42446100_78017.287_00006639
|
| 97 |
+
ARKitScenes_test Validation_42446100_78019.286_00006759
|
| 98 |
+
ARKitScenes_test Validation_42446100_78027.283_00007239
|
| 99 |
+
ARKitScenes_test Validation_42446100_78028.282_00007299
|
| 100 |
+
ARKitScenes_test Validation_42446103_78062.186_00000583
|
| 101 |
+
ARKitScenes_test Validation_42446103_78083.194_00001844
|
| 102 |
+
ARKitScenes_test Validation_42446103_78090.191_00002264
|
| 103 |
+
ARKitScenes_test Validation_42446103_78096.189_00002624
|
| 104 |
+
ARKitScenes_test Validation_42446103_78130.192_00004665
|
| 105 |
+
ARKitScenes_test Validation_42446103_78137.189_00005085
|
| 106 |
+
ARKitScenes_test Validation_42446103_78139.188_00005205
|
| 107 |
+
ARKitScenes_test Validation_42446103_78157.181_00006285
|
| 108 |
+
ARKitScenes_test Validation_42446103_78177.190_00007486
|
| 109 |
+
ARKitScenes_test Validation_42446114_78198.981_00000637
|
| 110 |
+
ARKitScenes_test Validation_42446114_78214.992_00001598
|
| 111 |
+
ARKitScenes_test Validation_42446114_78222.989_00002078
|
| 112 |
+
ARKitScenes_test Validation_42446114_78245.996_00003459
|
| 113 |
+
ARKitScenes_test Validation_42446114_78251.994_00003819
|
| 114 |
+
ARKitScenes_test Validation_42446114_78272.986_00005079
|
| 115 |
+
ARKitScenes_test Validation_42446114_78292.994_00006280
|
| 116 |
+
ARKitScenes_test Validation_42446114_78297.992_00006580
|
| 117 |
+
ARKitScenes_test Validation_42446114_78298.992_00006640
|
| 118 |
+
ARKitScenes_test Validation_42446114_78306.989_00007120
|
| 119 |
+
ARKitScenes_test Validation_42446114_78311.987_00007420
|
| 120 |
+
ARKitScenes_test Validation_42446114_78324.982_00008200
|
| 121 |
+
ARKitScenes_test Validation_42446114_78367.981_00010781
|
| 122 |
+
ARKitScenes_test Validation_42446156_79137.122_00000097
|
| 123 |
+
ARKitScenes_test Validation_42446156_79149.117_00000817
|
| 124 |
+
ARKitScenes_test Validation_42446156_79224.121_00005319
|
| 125 |
+
ARKitScenes_test Validation_42446163_79569.523_00000217
|
| 126 |
+
ARKitScenes_test Validation_42446163_79570.522_00000277
|
| 127 |
+
ARKitScenes_test Validation_42446163_79603.526_00002258
|
| 128 |
+
ARKitScenes_test Validation_42446163_79604.525_00002318
|
| 129 |
+
ARKitScenes_test Validation_42446163_79608.524_00002558
|
| 130 |
+
ARKitScenes_test Validation_42446163_79619.519_00003218
|
| 131 |
+
ARKitScenes_test Validation_42446163_79627.533_00003699
|
| 132 |
+
ARKitScenes_test Validation_42446165_79718.730_00000217
|
| 133 |
+
ARKitScenes_test Validation_42446165_79724.728_00000577
|
| 134 |
+
ARKitScenes_test Validation_42446165_79745.719_00001837
|
| 135 |
+
ARKitScenes_test Validation_42446165_79771.726_00003398
|
| 136 |
+
ARKitScenes_test Validation_42446165_79792.734_00004659
|
| 137 |
+
ARKitScenes_test Validation_42446167_79644.726_00000103
|
| 138 |
+
ARKitScenes_test Validation_42446167_79645.726_00000163
|
| 139 |
+
ARKitScenes_test Validation_42446167_79654.722_00000703
|
| 140 |
+
ARKitScenes_test Validation_42446167_79676.730_00002024
|
| 141 |
+
ARKitScenes_test Validation_42446167_79682.728_00002384
|
| 142 |
+
ARKitScenes_test Validation_42446167_79707.734_00003885
|
| 143 |
+
ARKitScenes_test Validation_42446517_197447.270_00001777
|
| 144 |
+
ARKitScenes_test Validation_42446519_197506.878_00001358
|
| 145 |
+
ARKitScenes_test Validation_42446522_197316.873_00000186
|
| 146 |
+
ARKitScenes_test Validation_42446522_197359.872_00002767
|
| 147 |
+
ARKitScenes_test Validation_42446522_197360.872_00002827
|
| 148 |
+
ARKitScenes_test Validation_42446522_197384.879_00004268
|
| 149 |
+
ARKitScenes_test Validation_42446522_197393.875_00004808
|
| 150 |
+
ARKitScenes_test Validation_42446522_197396.874_00004988
|
| 151 |
+
ARKitScenes_test Validation_42446527_198633.848_00000481
|
| 152 |
+
ARKitScenes_test Validation_42446527_198646.843_00001261
|
| 153 |
+
ARKitScenes_test Validation_42446527_198667.851_00002522
|
| 154 |
+
ARKitScenes_test Validation_42446527_198696.839_00004262
|
| 155 |
+
ARKitScenes_test Validation_42446529_198739.838_00001093
|
| 156 |
+
ARKitScenes_test Validation_42446529_198779.838_00003494
|
| 157 |
+
ARKitScenes_test Validation_42446532_198859.939_00003140
|
| 158 |
+
ARKitScenes_test Validation_42446532_198900.938_00005601
|
| 159 |
+
ARKitScenes_test Validation_42446533_200400.295_00000415
|
| 160 |
+
ARKitScenes_test Validation_42446533_200426.301_00001976
|
| 161 |
+
ARKitScenes_test Validation_42446536_200466.901_00001375
|
| 162 |
+
ARKitScenes_test Validation_42446536_200467.901_00001435
|
| 163 |
+
ARKitScenes_test Validation_42446540_201470.345_00001171
|
| 164 |
+
ARKitScenes_test Validation_42446540_201486.338_00002131
|
| 165 |
+
ARKitScenes_test Validation_42446540_201487.338_00002191
|
| 166 |
+
ARKitScenes_test Validation_42446540_201494.335_00002611
|
| 167 |
+
ARKitScenes_test Validation_42446540_201498.350_00002852
|
| 168 |
+
ARKitScenes_test Validation_42446540_201513.344_00003752
|
| 169 |
+
ARKitScenes_test Validation_42446540_201518.342_00004052
|
| 170 |
+
ARKitScenes_test Validation_42446540_201530.337_00004772
|
| 171 |
+
ARKitScenes_test Validation_42446541_201561.441_00000823
|
| 172 |
+
ARKitScenes_test Validation_42446541_201570.437_00001363
|
| 173 |
+
ARKitScenes_test Validation_42446541_201594.444_00002804
|
| 174 |
+
ARKitScenes_test Validation_42446541_201622.449_00004485
|
| 175 |
+
ARKitScenes_test Validation_42446541_201644.440_00005805
|
| 176 |
+
ARKitScenes_test Validation_42446541_201645.439_00005865
|
| 177 |
+
ARKitScenes_test Validation_42446541_201665.448_00007066
|
| 178 |
+
ARKitScenes_test Validation_42897501_270038.590_00001826
|
| 179 |
+
ARKitScenes_test Validation_42897501_270046.587_00002306
|
| 180 |
+
ARKitScenes_test Validation_42897501_270054.584_00002786
|
| 181 |
+
ARKitScenes_test Validation_42897501_270057.583_00002966
|
| 182 |
+
ARKitScenes_test Validation_42897501_270063.580_00003326
|
| 183 |
+
ARKitScenes_test Validation_42897504_269851.484_00000559
|
| 184 |
+
ARKitScenes_test Validation_42897504_269871.492_00001760
|
| 185 |
+
ARKitScenes_test Validation_42897504_269876.490_00002060
|
| 186 |
+
ARKitScenes_test Validation_42897504_269882.488_00002420
|
| 187 |
+
ARKitScenes_test Validation_42897508_269924.104_00000104
|
| 188 |
+
ARKitScenes_test Validation_42897508_269929.085_00000403
|
| 189 |
+
ARKitScenes_test Validation_42897508_269938.082_00000943
|
| 190 |
+
ARKitScenes_test Validation_42897508_269951.093_00001724
|
| 191 |
+
ARKitScenes_test Validation_42897508_269954.092_00001904
|
| 192 |
+
ARKitScenes_test Validation_42897508_269959.090_00002204
|
| 193 |
+
ARKitScenes_test Validation_42897508_269964.088_00002504
|
| 194 |
+
ARKitScenes_test Validation_42897508_269971.085_00002924
|
| 195 |
+
ARKitScenes_test Validation_42897521_273955.896_00000181
|
| 196 |
+
ARKitScenes_test Validation_42897521_273961.894_00000541
|
| 197 |
+
ARKitScenes_test Validation_42897521_273987.900_00002102
|
| 198 |
+
ARKitScenes_test Validation_42897526_273849.407_00000295
|
| 199 |
+
ARKitScenes_test Validation_42897526_273883.393_00002335
|
| 200 |
+
ARKitScenes_test Validation_42897526_273887.391_00002575
|
| 201 |
+
ARKitScenes_test Validation_42897526_273889.390_00002695
|
| 202 |
+
ARKitScenes_test Validation_42897528_273912.098_00000721
|
| 203 |
+
ARKitScenes_test Validation_42897528_273918.095_00001081
|
| 204 |
+
ARKitScenes_test Validation_42897542_275467.131_00000751
|
| 205 |
+
ARKitScenes_test Validation_42897542_275474.128_00001171
|
| 206 |
+
ARKitScenes_test Validation_42897542_275482.125_00001651
|
| 207 |
+
ARKitScenes_test Validation_42897545_276443.861_00000342
|
| 208 |
+
ARKitScenes_test Validation_42897549_276516.364_00001867
|
| 209 |
+
ARKitScenes_test Validation_42897550_276550.167_00001057
|
| 210 |
+
ARKitScenes_test Validation_42897552_277375.808_00001057
|
| 211 |
+
ARKitScenes_test Validation_42897559_277439.598_00000541
|
| 212 |
+
ARKitScenes_test Validation_42897559_277473.601_00002582
|
| 213 |
+
ARKitScenes_test Validation_42897559_277476.600_00002762
|
| 214 |
+
ARKitScenes_test Validation_42897559_277477.599_00002822
|
| 215 |
+
ARKitScenes_test Validation_42897561_278935.714_00001459
|
| 216 |
+
ARKitScenes_test Validation_42897561_278941.712_00001819
|
| 217 |
+
ARKitScenes_test Validation_42897561_278944.711_00001999
|
| 218 |
+
ARKitScenes_test Validation_42897561_278957.705_00002779
|
| 219 |
+
ARKitScenes_test Validation_42897564_278991.308_00001237
|
| 220 |
+
ARKitScenes_test Validation_42897564_279023.311_00003158
|
| 221 |
+
ARKitScenes_test Validation_42897566_279059.013_00001196
|
| 222 |
+
ARKitScenes_test Validation_42897566_279087.001_00002876
|
| 223 |
+
ARKitScenes_test Validation_42897599_476337.198_00002149
|
| 224 |
+
ARKitScenes_test Validation_42897599_476424.212_00007372
|
| 225 |
+
ARKitScenes_test Validation_42897599_476451.201_00008992
|
| 226 |
+
ARKitScenes_test Validation_42897599_476465.212_00009833
|
| 227 |
+
ARKitScenes_test Validation_42897599_476473.209_00010313
|
| 228 |
+
ARKitScenes_test Validation_42897599_476510.210_00012534
|
| 229 |
+
ARKitScenes_test Validation_42897599_476511.210_00012594
|
| 230 |
+
ARKitScenes_test Validation_42897599_476548.211_00014815
|
| 231 |
+
ARKitScenes_test Validation_42897599_476553.209_00015115
|
| 232 |
+
ARKitScenes_test Validation_42897599_476562.206_00015655
|
| 233 |
+
ARKitScenes_test Validation_42897647_478138.188_00000523
|
| 234 |
+
ARKitScenes_test Validation_42897647_478149.183_00001183
|
| 235 |
+
ARKitScenes_test Validation_42897647_478150.183_00001243
|
| 236 |
+
ARKitScenes_test Validation_42897647_478160.179_00001843
|
| 237 |
+
ARKitScenes_test Validation_42897647_478167.176_00002263
|
| 238 |
+
ARKitScenes_test Validation_42897647_478170.175_00002443
|
| 239 |
+
ARKitScenes_test Validation_42897647_478208.176_00004724
|
| 240 |
+
ARKitScenes_test Validation_42897647_478226.185_00005805
|
| 241 |
+
ARKitScenes_test Validation_42897651_478328.276_00005036
|
| 242 |
+
ARKitScenes_test Validation_42897651_478335.274_00005456
|
| 243 |
+
ARKitScenes_test Validation_42897651_478342.287_00005877
|
| 244 |
+
ARKitScenes_test Validation_42897651_478353.283_00006537
|
| 245 |
+
ARKitScenes_test Validation_42897651_478356.282_00006717
|
| 246 |
+
ARKitScenes_test Validation_42897651_478443.279_00011939
|
| 247 |
+
ARKitScenes_test Validation_42897651_478480.280_00014160
|
| 248 |
+
ARKitScenes_test Validation_42897651_478487.278_00014580
|
| 249 |
+
ARKitScenes_test Validation_42897651_478519.281_00016501
|
| 250 |
+
ARKitScenes_test Validation_42897651_478543.271_00017941
|
| 251 |
+
ARKitScenes_test Validation_42897667_478004.076_00000456
|
| 252 |
+
ARKitScenes_test Validation_42897667_478015.088_00001117
|
| 253 |
+
ARKitScenes_test Validation_42897667_478031.082_00002077
|
| 254 |
+
ARKitScenes_test Validation_42897667_478032.081_00002137
|
| 255 |
+
ARKitScenes_test Validation_42897667_478033.081_00002197
|
| 256 |
+
ARKitScenes_test Validation_42897667_478053.089_00003398
|
| 257 |
+
ARKitScenes_test Validation_42897667_478083.077_00005198
|
| 258 |
+
ARKitScenes_test Validation_42897667_478117.080_00007239
|
| 259 |
+
ARKitScenes_test Validation_42897688_479183.081_00002377
|
| 260 |
+
ARKitScenes_test Validation_42897688_479204.090_00003638
|
| 261 |
+
ARKitScenes_test Validation_42897688_479223.082_00004778
|
| 262 |
+
ARKitScenes_test Validation_42897688_479228.080_00005078
|
| 263 |
+
ARKitScenes_test Validation_42897688_479253.086_00006579
|
| 264 |
+
ARKitScenes_test Validation_42897688_479277.076_00008019
|
| 265 |
+
ARKitScenes_test Validation_42897688_479292.087_00008920
|
| 266 |
+
ARKitScenes_test Validation_42898811_167510.387_00002377
|
| 267 |
+
ARKitScenes_test Validation_42898816_167394.902_00002504
|
| 268 |
+
ARKitScenes_test Validation_42898818_168161.177_00000577
|
| 269 |
+
ARKitScenes_test Validation_42898818_168172.172_00001237
|
| 270 |
+
ARKitScenes_test Validation_42898818_168173.172_00001297
|
| 271 |
+
ARKitScenes_test Validation_42898822_168209.574_00000277
|
| 272 |
+
ARKitScenes_test Validation_42898822_168250.573_00002738
|
| 273 |
+
ARKitScenes_test Validation_42898826_168118.178_00001243
|
| 274 |
+
ARKitScenes_test Validation_42898826_168119.178_00001303
|
| 275 |
+
ARKitScenes_test Validation_42898849_196459.913_00000877
|
| 276 |
+
ARKitScenes_test Validation_42898849_196460.913_00000937
|
| 277 |
+
ARKitScenes_test Validation_42898849_196468.926_00001418
|
| 278 |
+
ARKitScenes_test Validation_42898849_196471.925_00001598
|
| 279 |
+
ARKitScenes_test Validation_42898849_196503.912_00003518
|
| 280 |
+
ARKitScenes_test Validation_42898849_196531.918_00005199
|
| 281 |
+
ARKitScenes_test Validation_42898849_196542.913_00005859
|
| 282 |
+
ARKitScenes_test Validation_42898849_196550.927_00006340
|
| 283 |
+
ARKitScenes_test Validation_42898849_196566.921_00007300
|
| 284 |
+
ARKitScenes_test Validation_42898849_196584.913_00008380
|
| 285 |
+
ARKitScenes_test Validation_42898854_196220.724_00002288
|
| 286 |
+
ARKitScenes_test Validation_42898854_196229.721_00002828
|
| 287 |
+
ARKitScenes_test Validation_42898854_196233.719_00003068
|
| 288 |
+
ARKitScenes_test Validation_42898854_196236.718_00003248
|
| 289 |
+
ARKitScenes_test Validation_42898854_196266.723_00005049
|
| 290 |
+
ARKitScenes_test Validation_42898854_196272.720_00005409
|
| 291 |
+
ARKitScenes_test Validation_42898854_196274.720_00005529
|
| 292 |
+
ARKitScenes_test Validation_42898854_196285.715_00006189
|
| 293 |
+
ARKitScenes_test Validation_42898862_196315.720_00000271
|
| 294 |
+
ARKitScenes_test Validation_42898862_196349.723_00002312
|
| 295 |
+
ARKitScenes_test Validation_42898862_196366.716_00003332
|
| 296 |
+
ARKitScenes_test Validation_42898862_196372.714_00003692
|
| 297 |
+
ARKitScenes_test Validation_42898862_196375.713_00003872
|
| 298 |
+
ARKitScenes_test Validation_42898862_196379.711_00004112
|
| 299 |
+
ARKitScenes_test Validation_42898862_196388.724_00004653
|
| 300 |
+
ARKitScenes_test Validation_42898862_196400.720_00005373
|
| 301 |
+
ARKitScenes_test Validation_42898862_196401.719_00005433
|
| 302 |
+
ARKitScenes_test Validation_42898862_196404.718_00005613
|
| 303 |
+
ARKitScenes_test Validation_42898862_196405.718_00005673
|
| 304 |
+
ARKitScenes_test Validation_42898862_196416.713_00006333
|
| 305 |
+
ARKitScenes_test Validation_42898862_196426.726_00006934
|
| 306 |
+
ARKitScenes_test Validation_42899461_195789.434_00000277
|
| 307 |
+
ARKitScenes_test Validation_42899461_195799.430_00000877
|
| 308 |
+
ARKitScenes_test Validation_42899461_195831.433_00002798
|
| 309 |
+
ARKitScenes_test Validation_42899461_195850.425_00003938
|
| 310 |
+
ARKitScenes_test Validation_42899461_195859.421_00004478
|
| 311 |
+
ARKitScenes_test Validation_42899461_195866.418_00004898
|
| 312 |
+
ARKitScenes_test Validation_42899471_195717.530_00003579
|
| 313 |
+
ARKitScenes_test Validation_42899471_195718.530_00003639
|
| 314 |
+
ARKitScenes_test Validation_42899611_228351.719_00000157
|
| 315 |
+
ARKitScenes_test Validation_42899611_228363.714_00000877
|
| 316 |
+
ARKitScenes_test Validation_42899611_228389.720_00002438
|
| 317 |
+
ARKitScenes_test Validation_42899611_228392.719_00002618
|
| 318 |
+
ARKitScenes_test Validation_42899611_228395.718_00002798
|
| 319 |
+
ARKitScenes_test Validation_42899612_228436.519_00000649
|
| 320 |
+
ARKitScenes_test Validation_42899612_228440.517_00000889
|
| 321 |
+
ARKitScenes_test Validation_42899612_228494.512_00004130
|
| 322 |
+
ARKitScenes_test Validation_42899612_228495.512_00004190
|
| 323 |
+
ARKitScenes_test Validation_42899612_228504.525_00004731
|
| 324 |
+
ARKitScenes_test Validation_42899612_228518.520_00005571
|
| 325 |
+
ARKitScenes_test Validation_42899613_228276.016_00000222
|
| 326 |
+
ARKitScenes_test Validation_42899613_228281.014_00000522
|
| 327 |
+
ARKitScenes_test Validation_42899613_228313.018_00002443
|
| 328 |
+
ARKitScenes_test Validation_42899613_228327.012_00003283
|
| 329 |
+
ARKitScenes_test Validation_42899619_228891.872_00000217
|
| 330 |
+
ARKitScenes_test Validation_42899619_228897.869_00000577
|
| 331 |
+
ARKitScenes_test Validation_42899620_228853.670_00000397
|
| 332 |
+
ARKitScenes_test Validation_42899620_228857.669_00000637
|
| 333 |
+
ARKitScenes_test Validation_44358442_42920.105_00001784
|
| 334 |
+
ARKitScenes_test Validation_44358446_43014.899_00000397
|
| 335 |
+
ARKitScenes_test Validation_44358446_43015.898_00000457
|
| 336 |
+
ARKitScenes_test Validation_44358446_43085.903_00004659
|
| 337 |
+
ARKitScenes_test Validation_44358446_43089.901_00004899
|
| 338 |
+
ARKitScenes_test Validation_44358446_43101.896_00005619
|
| 339 |
+
ARKitScenes_test Validation_44358446_43119.889_00006699
|
| 340 |
+
ARKitScenes_test Validation_44358448_42964.803_00000895
|
| 341 |
+
ARKitScenes_test Validation_44358451_43632.897_00001658
|
| 342 |
+
ARKitScenes_test Validation_44358451_43712.880_00006459
|
| 343 |
+
ARKitScenes_test Validation_44358452_43503.584_00000216
|
| 344 |
+
ARKitScenes_test Validation_44358452_43538.586_00002317
|
| 345 |
+
ARKitScenes_test Validation_44358455_43441.293_00002858
|
| 346 |
+
ARKitScenes_test Validation_44358455_43481.293_00005259
|
| 347 |
+
ARKitScenes_test Validation_45260854_561.091_00000289
|
| 348 |
+
ARKitScenes_test Validation_45260854_568.088_00000709
|
| 349 |
+
ARKitScenes_test Validation_45260854_570.087_00000829
|
| 350 |
+
ARKitScenes_test Validation_45260854_582.082_00001549
|
| 351 |
+
ARKitScenes_test Validation_45260856_640.292_00000877
|
| 352 |
+
ARKitScenes_test Validation_45260856_648.288_00001357
|
| 353 |
+
ARKitScenes_test Validation_45260856_653.286_00001657
|
| 354 |
+
ARKitScenes_test Validation_45260856_698.284_00004358
|
| 355 |
+
ARKitScenes_test Validation_45260857_519.692_00001358
|
| 356 |
+
ARKitScenes_test Validation_45260857_529.687_00001958
|
| 357 |
+
ARKitScenes_test Validation_45260857_534.685_00002258
|
| 358 |
+
ARKitScenes_test Validation_45260857_536.685_00002378
|
| 359 |
+
ARKitScenes_test Validation_45260857_539.683_00002558
|
| 360 |
+
ARKitScenes_test Validation_45260898_769.990_00000409
|
| 361 |
+
ARKitScenes_test Validation_45260898_774.988_00000709
|
| 362 |
+
ARKitScenes_test Validation_45260899_868.784_00001243
|
| 363 |
+
ARKitScenes_test Validation_45260899_885.794_00002264
|
| 364 |
+
ARKitScenes_test Validation_45260899_895.790_00002864
|
| 365 |
+
ARKitScenes_test Validation_45260900_804.493_00000073
|
| 366 |
+
ARKitScenes_test Validation_45260920_2471.417_00000288
|
| 367 |
+
ARKitScenes_test Validation_45260920_2474.416_00000468
|
| 368 |
+
ARKitScenes_test Validation_45260920_2494.425_00001669
|
| 369 |
+
ARKitScenes_test Validation_45260920_2512.417_00002749
|
| 370 |
+
ARKitScenes_test Validation_45260920_2535.425_00004130
|
| 371 |
+
ARKitScenes_test Validation_45260925_2412.823_00001117
|
| 372 |
+
ARKitScenes_test Validation_45260925_2424.819_00001837
|
| 373 |
+
ARKitScenes_test Validation_45260925_2427.818_00002017
|
| 374 |
+
ARKitScenes_test Validation_45260925_2454.824_00003638
|
| 375 |
+
ARKitScenes_test Validation_45260928_2345.217_00000457
|
| 376 |
+
ARKitScenes_test Validation_45260928_2362.227_00001478
|
| 377 |
+
ARKitScenes_test Validation_45260928_2371.223_00002018
|
| 378 |
+
ARKitScenes_test Validation_45261121_3447.063_00001994
|
| 379 |
+
ARKitScenes_test Validation_45261121_3448.063_00002054
|
| 380 |
+
ARKitScenes_test Validation_45261121_3454.060_00002414
|
| 381 |
+
ARKitScenes_test Validation_45261121_3468.071_00003255
|
| 382 |
+
ARKitScenes_test Validation_45261128_3526.164_00002198
|
| 383 |
+
ARKitScenes_test Validation_45261128_3536.159_00002798
|
| 384 |
+
ARKitScenes_test Validation_45261133_4030.272_00001417
|
| 385 |
+
ARKitScenes_test Validation_45261133_4040.268_00002017
|
| 386 |
+
ARKitScenes_test Validation_45261133_4047.265_00002437
|
| 387 |
+
ARKitScenes_test Validation_45261133_4048.265_00002497
|
| 388 |
+
ARKitScenes_test Validation_45261133_4050.264_00002617
|
| 389 |
+
ARKitScenes_test Validation_45261133_4079.268_00004358
|
| 390 |
+
ARKitScenes_test Validation_45261133_4080.268_00004418
|
| 391 |
+
ARKitScenes_test Validation_45261133_4107.273_00006039
|
| 392 |
+
ARKitScenes_test Validation_45261133_4116.270_00006579
|
| 393 |
+
ARKitScenes_test Validation_45261133_4117.269_00006639
|
| 394 |
+
ARKitScenes_test Validation_45261133_4126.266_00007179
|
| 395 |
+
ARKitScenes_test Validation_45261133_4132.263_00007539
|
| 396 |
+
ARKitScenes_test Validation_45261133_4136.261_00007779
|
| 397 |
+
ARKitScenes_test Validation_45261133_4142.259_00008139
|
| 398 |
+
ARKitScenes_test Validation_45261133_4148.256_00008499
|
| 399 |
+
ARKitScenes_test Validation_45261140_3951.272_00000283
|
| 400 |
+
ARKitScenes_test Validation_45261140_3960.268_00000823
|
| 401 |
+
ARKitScenes_test Validation_45261140_3976.261_00001783
|
| 402 |
+
ARKitScenes_test Validation_45261140_3978.260_00001903
|
| 403 |
+
ARKitScenes_test Validation_45261142_3888.464_00000174
|
| 404 |
+
ARKitScenes_test Validation_45261142_3896.461_00000654
|
| 405 |
+
ARKitScenes_test Validation_45261143_4588.703_00000583
|
| 406 |
+
ARKitScenes_test Validation_45261143_4592.701_00000823
|
| 407 |
+
ARKitScenes_test Validation_45261143_4628.703_00002984
|
| 408 |
+
ARKitScenes_test Validation_45261144_4658.007_00000457
|
| 409 |
+
ARKitScenes_test Validation_45261150_4531.310_00000337
|
| 410 |
+
ARKitScenes_test Validation_45261150_4535.308_00000577
|
| 411 |
+
ARKitScenes_test Validation_45261150_4549.302_00001417
|
| 412 |
+
ARKitScenes_test Validation_45261181_2852.026_00001537
|
| 413 |
+
ARKitScenes_test Validation_45261182_2977.726_00003818
|
| 414 |
+
ARKitScenes_test Validation_45261182_2994.736_00004839
|
| 415 |
+
ARKitScenes_test Validation_45261185_3558.482_00000703
|
| 416 |
+
ARKitScenes_test Validation_45261185_3559.481_00000763
|
| 417 |
+
ARKitScenes_test Validation_45261185_3562.480_00000943
|
| 418 |
+
ARKitScenes_test Validation_45261185_3601.481_00003284
|
| 419 |
+
ARKitScenes_test Validation_45261190_3716.886_00001843
|
| 420 |
+
ARKitScenes_test Validation_45261190_3743.892_00003464
|
| 421 |
+
ARKitScenes_test Validation_45261193_3625.189_00000409
|
| 422 |
+
ARKitScenes_test Validation_45261193_3632.186_00000829
|
| 423 |
+
ARKitScenes_test Validation_45261193_3655.193_00002210
|
| 424 |
+
ARKitScenes_test Validation_45261575_77474.508_00000337
|
| 425 |
+
ARKitScenes_test Validation_45261575_77475.507_00000397
|
| 426 |
+
ARKitScenes_test Validation_45261581_77518.506_00000157
|
| 427 |
+
ARKitScenes_test Validation_45261581_77535.516_00001178
|
| 428 |
+
ARKitScenes_test Validation_45261594_78039.207_00000637
|
| 429 |
+
ARKitScenes_test Validation_45261615_55008.692_00000703
|
| 430 |
+
ARKitScenes_test Validation_45261619_55088.693_00001436
|
| 431 |
+
ARKitScenes_test Validation_45261619_55116.682_00003116
|
| 432 |
+
ARKitScenes_test Validation_45261620_54934.288_00000157
|
| 433 |
+
ARKitScenes_test Validation_45261631_56030.404_00000217
|
| 434 |
+
ARKitScenes_test Validation_45261631_56068.389_00002497
|
| 435 |
+
ARKitScenes_test Validation_45261631_56089.397_00003758
|
| 436 |
+
ARKitScenes_test Validation_45261631_56092.396_00003938
|
| 437 |
+
ARKitScenes_test Validation_45261631_56104.391_00004658
|
| 438 |
+
ARKitScenes_test Validation_45261632_55892.192_00000138
|
| 439 |
+
ARKitScenes_test Validation_45261637_55961.498_00000337
|
| 440 |
+
ARKitScenes_test Validation_45261637_55962.497_00000397
|
| 441 |
+
ARKitScenes_test Validation_45261637_55971.494_00000937
|
| 442 |
+
ARKitScenes_test Validation_45662921_21784.929_00002048
|
| 443 |
+
ARKitScenes_test Validation_45662921_21787.928_00002228
|
| 444 |
+
ARKitScenes_test Validation_45662921_21791.926_00002468
|
| 445 |
+
ARKitScenes_test Validation_45662921_21836.924_00005169
|
| 446 |
+
ARKitScenes_test Validation_45662924_21592.925_00002258
|
| 447 |
+
ARKitScenes_test Validation_45662924_21607.919_00003158
|
| 448 |
+
ARKitScenes_test Validation_45662924_21609.918_00003278
|
| 449 |
+
ARKitScenes_test Validation_45662926_21649.319_00000667
|
| 450 |
+
ARKitScenes_test Validation_45662926_21659.331_00001268
|
| 451 |
+
ARKitScenes_test Validation_45662926_21670.326_00001928
|
| 452 |
+
ARKitScenes_test Validation_45662926_21674.325_00002168
|
| 453 |
+
ARKitScenes_test Validation_45662926_21675.324_00002228
|
| 454 |
+
ARKitScenes_test Validation_45662926_21682.322_00002648
|
| 455 |
+
ARKitScenes_test Validation_45662926_21684.321_00002768
|
| 456 |
+
ARKitScenes_test Validation_45662926_21688.319_00003008
|
| 457 |
+
ARKitScenes_test Validation_45662926_21692.317_00003248
|
| 458 |
+
ARKitScenes_test Validation_45662926_21693.317_00003308
|
| 459 |
+
ARKitScenes_test Validation_45662926_21694.317_00003368
|
| 460 |
+
ARKitScenes_test Validation_45662926_21711.326_00004389
|
| 461 |
+
ARKitScenes_test Validation_45662926_21715.324_00004629
|
| 462 |
+
ARKitScenes_test Validation_45662942_22918.660_00002197
|
| 463 |
+
ARKitScenes_test Validation_45662943_23032.263_00000817
|
| 464 |
+
ARKitScenes_test Validation_45662943_23164.258_00008740
|
| 465 |
+
ARKitScenes_test Validation_45662943_23168.256_00008980
|
| 466 |
+
ARKitScenes_test Validation_45662943_23170.255_00009100
|
| 467 |
+
ARKitScenes_test Validation_45662943_23171.255_00009160
|
| 468 |
+
ARKitScenes_test Validation_45662943_23174.254_00009340
|
| 469 |
+
ARKitScenes_test Validation_45662943_23175.253_00009400
|
| 470 |
+
ARKitScenes_test Validation_45662943_23179.268_00009641
|
| 471 |
+
ARKitScenes_test Validation_45662943_23183.267_00009881
|
| 472 |
+
ARKitScenes_test Validation_45662943_23208.256_00011381
|
| 473 |
+
ARKitScenes_test Validation_45662944_22963.758_00001297
|
| 474 |
+
ARKitScenes_test Validation_45662944_22974.770_00001958
|
| 475 |
+
ARKitScenes_test Validation_45662970_26159.531_00003188
|
| 476 |
+
ARKitScenes_test Validation_45662979_26048.725_00001177
|
| 477 |
+
ARKitScenes_test Validation_45662981_27937.116_00000050
|
| 478 |
+
ARKitScenes_test Validation_45662981_27946.096_00000589
|
| 479 |
+
ARKitScenes_test Validation_45662981_27994.094_00003470
|
| 480 |
+
ARKitScenes_test Validation_45662981_28029.097_00005571
|
| 481 |
+
ARKitScenes_test Validation_45662983_27801.886_00000337
|
| 482 |
+
ARKitScenes_test Validation_45662983_27805.885_00000577
|
| 483 |
+
ARKitScenes_test Validation_45662983_27807.884_00000697
|
| 484 |
+
ARKitScenes_test Validation_45662983_27851.883_00003338
|
| 485 |
+
ARKitScenes_test Validation_45662983_27854.882_00003518
|
| 486 |
+
ARKitScenes_test Validation_45662987_27869.193_00000169
|
| 487 |
+
ARKitScenes_test Validation_45662987_27878.190_00000709
|
| 488 |
+
ARKitScenes_test Validation_45662987_27883.188_00001009
|
| 489 |
+
ARKitScenes_test Validation_45662987_27910.194_00002630
|
| 490 |
+
ARKitScenes_test Validation_45662987_27911.193_00002690
|
| 491 |
+
ARKitScenes_test Validation_45663113_55308.661_00000277
|
| 492 |
+
ARKitScenes_test Validation_45663113_55324.671_00001238
|
| 493 |
+
ARKitScenes_test Validation_45663113_55326.671_00001358
|
| 494 |
+
ARKitScenes_test Validation_45663114_55386.862_00000637
|
| 495 |
+
ARKitScenes_test Validation_45663114_55404.855_00001717
|
| 496 |
+
ARKitScenes_test Validation_45663115_55461.865_00001958
|
| 497 |
+
ARKitScenes_test Validation_45663149_63304.709_00000517
|
| 498 |
+
ARKitScenes_test Validation_45663149_63311.706_00000937
|
| 499 |
+
ARKitScenes_test Validation_45663149_63314.705_00001117
|
| 500 |
+
ARKitScenes_test Validation_45663149_63315.705_00001177
|
| 501 |
+
ARKitScenes_test Validation_45663149_63324.701_00001717
|
| 502 |
+
ARKitScenes_test Validation_45663150_63511.952_00000288
|
| 503 |
+
ARKitScenes_test Validation_45663150_63551.953_00002689
|
| 504 |
+
ARKitScenes_test Validation_45663154_63399.880_00000074
|
| 505 |
+
ARKitScenes_test Validation_45663154_63401.862_00000193
|
| 506 |
+
ARKitScenes_test Validation_45663154_63409.859_00000673
|
| 507 |
+
ARKitScenes_test Validation_45663154_63411.858_00000793
|
| 508 |
+
ARKitScenes_test Validation_45663154_63450.860_00003134
|
| 509 |
+
ARKitScenes_test Validation_45663164_65947.828_00001015
|
| 510 |
+
ARKitScenes_test Validation_45663164_65948.828_00001075
|
| 511 |
+
ARKitScenes_test Validation_45663164_65968.820_00002275
|
| 512 |
+
ARKitScenes_test Validation_45663164_65975.817_00002695
|
| 513 |
+
ARKitScenes_test Validation_45663164_65997.825_00004016
|
| 514 |
+
ARKitScenes_test Validation_45663165_65815.014_00000637
|
| 515 |
+
ARKitScenes_test Validation_45663175_65890.618_00001153
|
| 516 |
+
ARKitScenes_test Validation_45663175_65891.617_00001213
|
| 517 |
+
ARKitScenes_test Validation_45663175_65912.625_00002474
|
| 518 |
+
ARKitScenes_test Validation_45663175_65919.623_00002894
|
| 519 |
+
ARKitScenes_test Validation_47115452_13057.568_00000097
|
| 520 |
+
ARKitScenes_test Validation_47115452_13076.576_00001238
|
| 521 |
+
ARKitScenes_test Validation_47115460_13145.781_00000529
|
| 522 |
+
ARKitScenes_test Validation_47115460_13166.772_00001789
|
| 523 |
+
ARKitScenes_test Validation_47115463_13220.467_00000096
|
| 524 |
+
ARKitScenes_test Validation_47115463_13235.477_00000997
|
| 525 |
+
ARKitScenes_test Validation_47115463_13282.474_00003818
|
| 526 |
+
ARKitScenes_test Validation_47115469_14055.292_00002678
|
| 527 |
+
ARKitScenes_test Validation_47115469_14059.290_00002918
|
| 528 |
+
ARKitScenes_test Validation_47115473_14107.787_00002150
|
| 529 |
+
ARKitScenes_test Validation_47115473_14172.776_00006051
|
| 530 |
+
ARKitScenes_test Validation_47115474_13966.678_00000402
|
| 531 |
+
ARKitScenes_test Validation_47115474_13976.691_00001003
|
| 532 |
+
ARKitScenes_test Validation_47204552_3357.092_00000277
|
| 533 |
+
ARKitScenes_test Validation_47204552_3378.100_00001538
|
| 534 |
+
ARKitScenes_test Validation_47204552_3391.095_00002318
|
| 535 |
+
ARKitScenes_test Validation_47204552_3412.086_00003578
|
| 536 |
+
ARKitScenes_test Validation_47204552_3425.097_00004359
|
| 537 |
+
ARKitScenes_test Validation_47204552_3430.095_00004659
|
| 538 |
+
ARKitScenes_test Validation_47204552_3441.091_00005319
|
| 539 |
+
ARKitScenes_test Validation_47204552_3485.089_00007960
|
| 540 |
+
ARKitScenes_test Validation_47204552_3486.089_00008020
|
| 541 |
+
ARKitScenes_test Validation_47204554_3215.201_00000277
|
| 542 |
+
ARKitScenes_test Validation_47204554_3227.196_00000997
|
| 543 |
+
ARKitScenes_test Validation_47204554_3240.191_00001777
|
| 544 |
+
ARKitScenes_test Validation_47204554_3244.189_00002017
|
| 545 |
+
ARKitScenes_test Validation_47204554_3261.199_00003038
|
| 546 |
+
ARKitScenes_test Validation_47204554_3263.198_00003158
|
| 547 |
+
ARKitScenes_test Validation_47204554_3265.197_00003278
|
| 548 |
+
ARKitScenes_test Validation_47204554_3269.195_00003518
|
| 549 |
+
ARKitScenes_test Validation_47204556_3286.405_00000037
|
| 550 |
+
ARKitScenes_test Validation_47204556_3312.394_00001597
|
| 551 |
+
ARKitScenes_test Validation_47204559_3992.536_00000276
|
| 552 |
+
ARKitScenes_test Validation_47204559_4001.549_00000817
|
| 553 |
+
ARKitScenes_test Validation_47204563_4044.448_00000697
|
| 554 |
+
ARKitScenes_test Validation_47204566_4109.038_00002017
|
| 555 |
+
ARKitScenes_test Validation_47204566_4118.051_00002558
|
| 556 |
+
ARKitScenes_test Validation_47204573_5014.617_00003459
|
| 557 |
+
ARKitScenes_test Validation_47204573_5039.606_00004959
|
| 558 |
+
ARKitScenes_test Validation_47204573_5046.620_00005380
|
| 559 |
+
ARKitScenes_test Validation_47204578_4918.223_00001484
|
| 560 |
+
ARKitScenes_test Validation_47204605_2434.303_00001922
|
| 561 |
+
ARKitScenes_test Validation_47204607_2521.802_00000097
|
| 562 |
+
ARKitScenes_test Validation_47204607_2522.802_00000157
|
| 563 |
+
ARKitScenes_test Validation_47204607_2523.801_00000217
|
| 564 |
+
ARKitScenes_test Validation_47204607_2554.806_00002078
|
| 565 |
+
ARKitScenes_test Validation_47204607_2590.791_00004238
|
| 566 |
+
ARKitScenes_test Validation_47204607_2635.790_00006939
|
| 567 |
+
ARKitScenes_test Validation_47204607_2676.791_00009400
|
| 568 |
+
ARKitScenes_test Validation_47204607_2677.790_00009460
|
| 569 |
+
ARKitScenes_test Validation_47204609_2459.993_00000457
|
| 570 |
+
ARKitScenes_test Validation_47204609_2462.992_00000637
|
| 571 |
+
ARKitScenes_test Validation_47331068_131079.065_00000877
|
| 572 |
+
ARKitScenes_test Validation_47331071_131004.361_00000637
|
| 573 |
+
ARKitScenes_test Validation_47331265_90018.298_00000433
|
| 574 |
+
ARKitScenes_test Validation_47331265_90026.295_00000913
|
| 575 |
+
ARKitScenes_test Validation_47331266_89826.492_00000319
|
| 576 |
+
ARKitScenes_test Validation_47331266_89870.491_00002960
|
| 577 |
+
ARKitScenes_test Validation_47331266_89893.499_00004341
|
| 578 |
+
ARKitScenes_test Validation_47331319_179047.025_00000654
|
| 579 |
+
ARKitScenes_test Validation_47331322_179161.329_00002870
|
| 580 |
+
ARKitScenes_test Validation_47331337_183502.425_00001897
|
| 581 |
+
ARKitScenes_test Validation_47331339_183421.825_00001579
|
| 582 |
+
ARKitScenes_test Validation_47331339_183426.823_00001879
|
| 583 |
+
ARKitScenes_test Validation_47331651_22884.238_00002078
|
| 584 |
+
ARKitScenes_test Validation_47331651_22902.231_00003158
|
| 585 |
+
ARKitScenes_test Validation_47331651_22903.230_00003218
|
| 586 |
+
ARKitScenes_test Validation_47331651_22905.230_00003338
|
| 587 |
+
ARKitScenes_test Validation_47331651_22921.223_00004298
|
| 588 |
+
ARKitScenes_test Validation_47331651_22924.239_00004479
|
| 589 |
+
ARKitScenes_test Validation_47331651_22928.237_00004719
|
| 590 |
+
ARKitScenes_test Validation_47331651_22930.236_00004839
|
| 591 |
+
ARKitScenes_test Validation_47331651_22939.232_00005379
|
| 592 |
+
ARKitScenes_test Validation_47331651_22947.229_00005859
|
| 593 |
+
ARKitScenes_test Validation_47331651_22971.236_00007300
|
| 594 |
+
ARKitScenes_test Validation_47331651_22972.236_00007360
|
| 595 |
+
ARKitScenes_test Validation_47331653_22994.327_00000577
|
| 596 |
+
ARKitScenes_test Validation_47331653_23001.324_00000997
|
| 597 |
+
ARKitScenes_test Validation_47331653_23054.335_00004179
|
| 598 |
+
ARKitScenes_test Validation_47331653_23064.331_00004779
|
| 599 |
+
ARKitScenes_test Validation_47331653_23111.329_00007600
|
| 600 |
+
ARKitScenes_test Validation_47331653_23114.328_00007780
|
| 601 |
+
ARKitScenes_test Validation_47331653_23127.322_00008560
|
| 602 |
+
ARKitScenes_test Validation_47331653_23133.336_00008921
|
| 603 |
+
ARKitScenes_test Validation_47331653_23134.336_00008981
|
| 604 |
+
ARKitScenes_test Validation_47331654_22743.829_00000457
|
| 605 |
+
ARKitScenes_test Validation_47331654_22785.828_00002978
|
| 606 |
+
ARKitScenes_test Validation_47331654_22786.828_00003038
|
| 607 |
+
ARKitScenes_test Validation_47331654_22832.826_00005799
|
| 608 |
+
ARKitScenes_test Validation_47331661_23633.580_00002870
|
| 609 |
+
ARKitScenes_test Validation_47331661_23643.593_00003471
|
| 610 |
+
ARKitScenes_test Validation_47331661_23644.592_00003531
|
| 611 |
+
ARKitScenes_test Validation_47331661_23685.592_00005992
|
| 612 |
+
ARKitScenes_test Validation_47331661_23688.591_00006172
|
| 613 |
+
ARKitScenes_test Validation_47331661_23699.587_00006832
|
| 614 |
+
ARKitScenes_test Validation_47331661_23720.578_00008092
|
| 615 |
+
ARKitScenes_test Validation_47331662_23462.083_00002497
|
| 616 |
+
ARKitScenes_test Validation_47331662_23468.081_00002857
|
| 617 |
+
ARKitScenes_test Validation_47331662_23469.080_00002917
|
| 618 |
+
ARKitScenes_test Validation_47331662_23485.091_00003878
|
| 619 |
+
ARKitScenes_test Validation_47331662_23488.089_00004058
|
| 620 |
+
ARKitScenes_test Validation_47331988_170632.558_00002264
|
| 621 |
+
ARKitScenes_test Validation_47331988_170653.550_00003524
|
| 622 |
+
ARKitScenes_test Validation_47331988_170659.548_00003884
|
| 623 |
+
ARKitScenes_test Validation_47331988_170749.561_00009287
|
| 624 |
+
ARKitScenes_test Validation_47331989_170421.860_00000457
|
| 625 |
+
ARKitScenes_test Validation_47331989_170455.863_00002498
|
| 626 |
+
ARKitScenes_test Validation_47331990_170521.953_00002509
|
| 627 |
+
ARKitScenes_test Validation_47331990_170525.952_00002749
|
| 628 |
+
ARKitScenes_test Validation_47331990_170549.959_00004190
|
| 629 |
+
ARKitScenes_test Validation_47331990_170558.955_00004730
|
| 630 |
+
ARKitScenes_test Validation_47332005_172103.731_00002443
|
| 631 |
+
ARKitScenes_test Validation_47332005_172170.737_00006465
|
| 632 |
+
ARKitScenes_test Validation_47332885_7493.862_00000583
|
| 633 |
+
ARKitScenes_test Validation_47332886_7578.261_00002215
|
| 634 |
+
ARKitScenes_test Validation_47332890_7645.367_00002918
|
| 635 |
+
ARKitScenes_test Validation_47332890_7647.366_00003038
|
| 636 |
+
ARKitScenes_test Validation_47332890_7669.357_00004358
|
| 637 |
+
ARKitScenes_test Validation_47332895_9056.792_00000655
|
| 638 |
+
ARKitScenes_test Validation_47332899_9178.193_00003548
|
| 639 |
+
ARKitScenes_test Validation_47332901_11035.593_00001165
|
| 640 |
+
ARKitScenes_test Validation_47332901_11041.590_00001525
|
| 641 |
+
ARKitScenes_test Validation_47332901_11042.590_00001585
|
| 642 |
+
ARKitScenes_test Validation_47332904_11139.400_00003705
|
| 643 |
+
ARKitScenes_test Validation_47332904_11147.397_00004185
|
| 644 |
+
ARKitScenes_test Validation_47332905_10983.897_00002030
|
| 645 |
+
ARKitScenes_test Validation_47332910_12199.324_00000768
|
| 646 |
+
ARKitScenes_test Validation_47332910_12253.336_00004010
|
| 647 |
+
ARKitScenes_test Validation_47332911_12284.823_00001543
|
| 648 |
+
ARKitScenes_test Validation_47332911_12350.829_00005505
|
| 649 |
+
ARKitScenes_test Validation_47332915_13471.971_00000343
|
| 650 |
+
ARKitScenes_test Validation_47332915_13507.973_00002504
|
| 651 |
+
ARKitScenes_test Validation_47332915_13543.975_00004665
|
| 652 |
+
ARKitScenes_test Validation_47332916_13398.568_00000475
|
| 653 |
+
ARKitScenes_test Validation_47332918_13325.864_00000589
|
| 654 |
+
ARKitScenes_test Validation_47332918_13336.876_00001250
|
| 655 |
+
ARKitScenes_test Validation_47332918_13368.863_00003170
|
| 656 |
+
ARKitScenes_test Validation_47333440_53804.226_00000637
|
| 657 |
+
ARKitScenes_test Validation_47333440_53839.228_00002738
|
| 658 |
+
ARKitScenes_test Validation_47333441_53869.716_00001117
|
| 659 |
+
ARKitScenes_test Validation_47333443_53783.118_00002077
|
| 660 |
+
ARKitScenes_test Validation_47333443_53784.117_00002137
|
| 661 |
+
ARKitScenes_test Validation_47333452_54954.122_00000157
|
| 662 |
+
ARKitScenes_test Validation_47333452_54987.208_00002143
|
| 663 |
+
ARKitScenes_test Validation_47333456_55036.122_00001718
|
| 664 |
+
ARKitScenes_test Validation_47333898_81373.971_00003278
|
| 665 |
+
ARKitScenes_test Validation_47333899_81292.071_00003098
|
| 666 |
+
ARKitScenes_test Validation_47333904_81413.372_00000697
|
| 667 |
+
ARKitScenes_test Validation_47333904_81420.369_00001117
|
| 668 |
+
ARKitScenes_test Validation_47333904_81424.367_00001357
|
| 669 |
+
ARKitScenes_test Validation_47333904_81426.367_00001477
|
| 670 |
+
ARKitScenes_test Validation_47333904_81435.363_00002017
|
| 671 |
+
ARKitScenes_test Validation_47333904_81443.376_00002498
|
| 672 |
+
ARKitScenes_test Validation_47333904_81444.376_00002558
|
| 673 |
+
ARKitScenes_test Validation_47333916_82429.566_00003699
|
| 674 |
+
ARKitScenes_test Validation_47333916_82433.564_00003939
|
| 675 |
+
ARKitScenes_test Validation_47333918_82508.450_00003163
|
| 676 |
+
ARKitScenes_test Validation_47333918_82513.465_00003464
|
| 677 |
+
ARKitScenes_test Validation_47333920_82646.261_00006096
|
| 678 |
+
ARKitScenes_test Validation_47333920_82649.260_00006276
|
| 679 |
+
ARKitScenes_test Validation_47333920_82658.256_00006816
|
| 680 |
+
ARKitScenes_test Validation_47333920_82663.254_00007116
|
| 681 |
+
ARKitScenes_test Validation_47333923_83608.499_00000636
|
| 682 |
+
ARKitScenes_test Validation_47333923_83645.501_00002857
|
| 683 |
+
ARKitScenes_test Validation_47333924_83693.915_00002078
|
| 684 |
+
ARKitScenes_test Validation_47333925_83549.606_00000523
|
| 685 |
+
ARKitScenes_test Validation_47333925_83550.606_00000583
|
| 686 |
+
ARKitScenes_test Validation_47333925_83581.610_00002444
|
| 687 |
+
ARKitScenes_test Validation_47333927_84003.213_00000824
|
| 688 |
+
ARKitScenes_test Validation_47333927_84012.209_00001364
|
| 689 |
+
ARKitScenes_test Validation_47333931_83942.704_00001778
|
| 690 |
+
ARKitScenes_test Validation_47333932_84111.202_00002918
|
| 691 |
+
ARKitScenes_test Validation_47333932_84112.202_00002978
|
| 692 |
+
ARKitScenes_test Validation_47333932_84125.196_00003758
|
| 693 |
+
ARKitScenes_test Validation_47333932_84183.206_00007240
|
| 694 |
+
ARKitScenes_test Validation_47333934_85569.766_00001249
|
| 695 |
+
ARKitScenes_test Validation_47333934_85621.762_00004370
|
| 696 |
+
ARKitScenes_test Validation_47333934_85645.769_00005811
|
| 697 |
+
ARKitScenes_test Validation_47333934_85655.765_00006411
|
| 698 |
+
ARKitScenes_test Validation_47333934_85660.763_00006711
|
| 699 |
+
ARKitScenes_test Validation_47333937_85506.376_00001021
|
| 700 |
+
ARKitScenes_test Validation_47333937_85512.373_00001381
|
| 701 |
+
ARKitScenes_test Validation_47333940_85460.977_00003038
|
| 702 |
+
ARKitScenes_test Validation_47333940_85482.968_00004358
|
| 703 |
+
ARKitScenes_test Validation_47334105_92895.748_00000769
|
| 704 |
+
ARKitScenes_test Validation_47334105_92897.747_00000889
|
| 705 |
+
ARKitScenes_test Validation_47334105_92903.744_00001249
|
| 706 |
+
ARKitScenes_test Validation_47334105_92922.753_00002390
|
| 707 |
+
ARKitScenes_test Validation_47334105_92946.744_00003830
|
| 708 |
+
ARKitScenes_test Validation_47334105_92951.758_00004131
|
| 709 |
+
ARKitScenes_test Validation_47334105_92954.757_00004311
|
| 710 |
+
ARKitScenes_test Validation_47334105_92986.744_00006231
|
| 711 |
+
ARKitScenes_test Validation_47334105_92990.759_00006472
|
| 712 |
+
ARKitScenes_test Validation_47334105_93022.746_00008392
|
| 713 |
+
ARKitScenes_test Validation_47334105_93029.743_00008812
|
| 714 |
+
ARKitScenes_test Validation_47334239_111112.020_00000517
|
| 715 |
+
ARKitScenes_test Validation_47334239_111149.022_00002738
|
| 716 |
+
ARKitScenes_test Validation_47334240_111193.421_00002017
|
| 717 |
+
ARKitScenes_test Validation_47334240_111214.429_00003278
|
| 718 |
+
ARKitScenes_test Validation_47334256_112032.695_00002258
|
| 719 |
+
ARKitScenes_test Validation_47334256_112063.700_00004119
|
| 720 |
+
ARKitScenes_test Validation_47334256_112064.699_00004179
|
| 721 |
+
ARKitScenes_test Validation_47334256_112065.699_00004239
|
| 722 |
+
ARKitScenes_test Validation_47429922_136624.256_00000517
|
| 723 |
+
ARKitScenes_test Validation_47429922_136657.242_00002497
|
| 724 |
+
ARKitScenes_test Validation_47429971_16349.760_00004238
|
| 725 |
+
ARKitScenes_test Validation_47430002_17171.758_00000277
|
| 726 |
+
ARKitScenes_test Validation_47430005_17517.360_00000276
|
| 727 |
+
ARKitScenes_test Validation_47430005_17532.371_00001177
|
| 728 |
+
ARKitScenes_test Validation_47430023_18906.937_00002864
|
| 729 |
+
ARKitScenes_test Validation_47430033_19414.387_00000403
|
| 730 |
+
ARKitScenes_test Validation_47430038_20536.784_00000216
|
| 731 |
+
ARKitScenes_test Validation_47430043_20606.288_00001717
|
| 732 |
+
ARKitScenes_test Validation_47430047_21122.842_00001898
|
| 733 |
+
ARKitScenes_test Validation_47895341_96403.324_00001057
|
| 734 |
+
ARKitScenes_test Validation_47895341_96411.321_00001537
|
| 735 |
+
ARKitScenes_test Validation_47895348_96488.523_00001777
|
| 736 |
+
ARKitScenes_test Validation_47895348_96538.520_00004778
|
| 737 |
+
ARKitScenes_test Validation_47895350_96557.812_00000396
|
| 738 |
+
ARKitScenes_test Validation_47895350_96559.828_00000517
|
| 739 |
+
ARKitScenes_test Validation_47895350_96571.823_00001237
|
| 740 |
+
ARKitScenes_test Validation_47895350_96604.826_00003218
|
| 741 |
+
ARKitScenes_test Validation_47895350_96610.824_00003578
|
| 742 |
+
ARKitScenes_test Validation_47895350_96621.819_00004238
|
| 743 |
+
ARKitScenes_test Validation_47895350_96646.826_00005739
|
| 744 |
+
ARKitScenes_test Validation_47895350_96659.821_00006519
|
| 745 |
+
ARKitScenes_test Validation_47895350_96663.819_00006759
|
| 746 |
+
ARKitScenes_test Validation_47895350_96730.825_00010781
|
| 747 |
+
ARKitScenes_test Validation_47895355_98775.628_00001297
|
| 748 |
+
ARKitScenes_test Validation_47895355_98788.639_00002078
|
| 749 |
+
ARKitScenes_test Validation_47895355_98796.636_00002558
|
| 750 |
+
ARKitScenes_test Validation_47895355_98810.630_00003398
|
| 751 |
+
ARKitScenes_test Validation_47895355_98823.642_00004179
|
| 752 |
+
ARKitScenes_test Validation_47895355_98827.640_00004419
|
| 753 |
+
ARKitScenes_test Validation_47895355_98906.624_00009160
|
| 754 |
+
ARKitScenes_test Validation_47895355_98910.639_00009401
|
| 755 |
+
ARKitScenes_test Validation_47895355_98929.632_00010541
|
| 756 |
+
ARKitScenes_test Validation_47895355_98943.626_00011381
|
| 757 |
+
ARKitScenes_test Validation_47895355_98966.633_00012762
|
| 758 |
+
ARKitScenes_test Validation_47895364_100221.941_00003158
|
| 759 |
+
ARKitScenes_test Validation_47895364_100251.946_00004959
|
| 760 |
+
ARKitScenes_test Validation_47895364_100287.931_00007119
|
| 761 |
+
ARKitScenes_test Validation_47895364_100307.940_00008320
|
| 762 |
+
ARKitScenes_test Validation_48018367_389452.575_00000457
|
| 763 |
+
ARKitScenes_test Validation_48018367_389473.583_00001718
|
| 764 |
+
ARKitScenes_test Validation_48018368_389553.283_00001190
|
| 765 |
+
ARKitScenes_test Validation_48018368_389572.275_00002330
|
| 766 |
+
ARKitScenes_test Validation_48018368_389581.271_00002870
|
| 767 |
+
ARKitScenes_test Validation_48018372_389531.675_00002498
|
| 768 |
+
ARKitScenes_test Validation_48018379_390040.136_00002678
|
| 769 |
+
ARKitScenes_test Validation_48018559_3346.848_00000337
|
| 770 |
+
ARKitScenes_test Validation_48018566_3858.438_00000103
|
| 771 |
+
ARKitScenes_test Validation_48018571_3905.036_00000439
|
| 772 |
+
ARKitScenes_test Validation_48018572_3798.046_00000128
|
| 773 |
+
ARKitScenes_test Validation_48018572_3810.041_00000848
|
| 774 |
+
ARKitScenes_test Validation_48018730_1678.470_00000637
|
| 775 |
+
ARKitScenes_test Validation_48018730_1723.469_00003338
|
| 776 |
+
ARKitScenes_test Validation_48018730_1769.467_00006099
|
| 777 |
+
ARKitScenes_test Validation_48018732_1939.564_00001759
|
| 778 |
+
ARKitScenes_test Validation_48018732_1942.563_00001939
|
| 779 |
+
ARKitScenes_test Validation_48018732_1945.579_00002120
|
| 780 |
+
ARKitScenes_test Validation_48018732_1948.577_00002300
|
| 781 |
+
ARKitScenes_test Validation_48018732_1953.575_00002600
|
| 782 |
+
ARKitScenes_test Validation_48018741_3393.825_00000337
|
| 783 |
+
ARKitScenes_test Validation_48458481_111522.301_00001117
|
| 784 |
+
ARKitScenes_test Validation_48458484_111590.308_00001237
|
| 785 |
+
ARKitScenes_test Validation_48458489_111549.307_00000703
|
| 786 |
+
ARKitScenes_test Validation_48458489_111557.304_00001183
|
| 787 |
+
ARKitScenes_test Validation_48458650_7066.602_00000456
|
| 788 |
+
ARKitScenes_test Validation_48458650_7068.601_00000576
|
| 789 |
+
ARKitScenes_test Validation_48458650_7070.617_00000697
|
| 790 |
+
ARKitScenes_test Validation_48458650_7071.617_00000757
|
| 791 |
+
ARKitScenes_test Validation_48458656_7661.589_00002077
|
| 792 |
+
ARKitScenes_test Validation_48458656_7663.589_00002197
|
| 793 |
+
ARKitScenes_test Validation_48458656_7667.587_00002437
|
| 794 |
+
ARKitScenes_test Validation_48458656_7672.585_00002737
|
| 795 |
+
ARKitScenes_test Validation_48458657_7617.691_00002023
|
| 796 |
+
ARKitScenes_test Validation_48458660_8270.961_00001658
|
| 797 |
+
ARKitScenes_test Validation_48458667_8365.655_00000637
|
| 798 |
+
ARKitScenes_test Validation_48458667_8367.654_00000757
|
| 799 |
+
ARKitScenes_test Validation_48458667_8376.651_00001297
|
| 800 |
+
ARKitScenes_test Validation_48458667_8395.660_00002438
|
external/WildCamera/splits/biwirgbdid_test.txt
ADDED
|
@@ -0,0 +1,800 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
BIWIRGBDID Testing_Still_000_000a_000018-a_111628807_rgb
|
| 2 |
+
BIWIRGBDID Testing_Still_000_000a_000019-a_111628891_rgb
|
| 3 |
+
BIWIRGBDID Testing_Still_000_000a_000124-a_111639723_rgb
|
| 4 |
+
BIWIRGBDID Testing_Still_000_000a_000157-a_111643139_rgb
|
| 5 |
+
BIWIRGBDID Testing_Still_000_000a_000179-a_111645307_rgb
|
| 6 |
+
BIWIRGBDID Testing_Still_000_000a_000183-a_111645807_rgb
|
| 7 |
+
BIWIRGBDID Testing_Still_000_000a_000195-a_111647059_rgb
|
| 8 |
+
BIWIRGBDID Testing_Still_000_000a_000200-a_111647559_rgb
|
| 9 |
+
BIWIRGBDID Testing_Still_001_001a_000022-a_106257171_rgb
|
| 10 |
+
BIWIRGBDID Testing_Still_001_001a_000039-a_106258923_rgb
|
| 11 |
+
BIWIRGBDID Testing_Still_001_001a_000056-a_106260671_rgb
|
| 12 |
+
BIWIRGBDID Testing_Still_001_001a_000090-a_106264423_rgb
|
| 13 |
+
BIWIRGBDID Testing_Still_001_001a_000099-a_106265423_rgb
|
| 14 |
+
BIWIRGBDID Testing_Still_001_001a_000117-a_106267255_rgb
|
| 15 |
+
BIWIRGBDID Testing_Still_001_001a_000145-a_106270255_rgb
|
| 16 |
+
BIWIRGBDID Testing_Still_001_001a_000157-a_106271503_rgb
|
| 17 |
+
BIWIRGBDID Testing_Still_002_002a_000016-a_1232924843_rgb
|
| 18 |
+
BIWIRGBDID Testing_Still_002_002a_000025-a_1232925843_rgb
|
| 19 |
+
BIWIRGBDID Testing_Still_002_002a_000035-a_1232926927_rgb
|
| 20 |
+
BIWIRGBDID Testing_Still_002_002a_000069-a_1232930595_rgb
|
| 21 |
+
BIWIRGBDID Testing_Still_002_002a_000088-a_1232932679_rgb
|
| 22 |
+
BIWIRGBDID Testing_Still_002_002a_000089-a_1232932759_rgb
|
| 23 |
+
BIWIRGBDID Testing_Still_002_002a_000092-a_1232933095_rgb
|
| 24 |
+
BIWIRGBDID Testing_Still_002_002a_000094-a_1232933259_rgb
|
| 25 |
+
BIWIRGBDID Testing_Still_002_002a_000123-a_1232936595_rgb
|
| 26 |
+
BIWIRGBDID Testing_Still_002_002a_000130-a_1232937342_rgb
|
| 27 |
+
BIWIRGBDID Testing_Still_002_002a_000141-a_1232938510_rgb
|
| 28 |
+
BIWIRGBDID Testing_Still_002_002a_000143-a_1232938759_rgb
|
| 29 |
+
BIWIRGBDID Testing_Still_002_002a_000146-a_1232939010_rgb
|
| 30 |
+
BIWIRGBDID Testing_Still_003_003a_000006-a_621308413_rgb
|
| 31 |
+
BIWIRGBDID Testing_Still_003_003a_000007-a_621308581_rgb
|
| 32 |
+
BIWIRGBDID Testing_Still_003_003a_000033-a_621311497_rgb
|
| 33 |
+
BIWIRGBDID Testing_Still_003_003a_000050-a_621313413_rgb
|
| 34 |
+
BIWIRGBDID Testing_Still_003_003a_000110-a_621320413_rgb
|
| 35 |
+
BIWIRGBDID Testing_Still_003_003a_000141-a_621324749_rgb
|
| 36 |
+
BIWIRGBDID Testing_Still_006_006a_000000-a_1224024579_rgb
|
| 37 |
+
BIWIRGBDID Testing_Still_006_006a_000002-a_1224024747_rgb
|
| 38 |
+
BIWIRGBDID Testing_Still_006_006a_000034-a_1224028080_rgb
|
| 39 |
+
BIWIRGBDID Testing_Still_006_006a_000055-a_1224030247_rgb
|
| 40 |
+
BIWIRGBDID Testing_Still_006_006a_000070-a_1224031831_rgb
|
| 41 |
+
BIWIRGBDID Testing_Still_006_006a_000089-a_1224033747_rgb
|
| 42 |
+
BIWIRGBDID Testing_Still_006_006a_000092-a_1224034079_rgb
|
| 43 |
+
BIWIRGBDID Testing_Still_006_006a_000098-a_1224034663_rgb
|
| 44 |
+
BIWIRGBDID Testing_Still_006_006a_000106-a_1224035495_rgb
|
| 45 |
+
BIWIRGBDID Testing_Still_006_006a_000122-a_1224037163_rgb
|
| 46 |
+
BIWIRGBDID Testing_Still_006_006a_000124-a_1224037331_rgb
|
| 47 |
+
BIWIRGBDID Testing_Still_006_006a_000139-a_1224038911_rgb
|
| 48 |
+
BIWIRGBDID Testing_Still_006_006a_000151-a_1224040163_rgb
|
| 49 |
+
BIWIRGBDID Testing_Still_008_008a_000014-a_190303117_rgb
|
| 50 |
+
BIWIRGBDID Testing_Still_008_008a_000029-a_190304618_rgb
|
| 51 |
+
BIWIRGBDID Testing_Still_008_008a_000073-a_190309201_rgb
|
| 52 |
+
BIWIRGBDID Testing_Still_008_008a_000092-a_190311285_rgb
|
| 53 |
+
BIWIRGBDID Testing_Still_008_008a_000116-a_190313701_rgb
|
| 54 |
+
BIWIRGBDID Testing_Still_008_008a_000125-a_190314617_rgb
|
| 55 |
+
BIWIRGBDID Testing_Still_010_010a_000005-a_171989565_rgb
|
| 56 |
+
BIWIRGBDID Testing_Still_010_010a_000019-a_171990981_rgb
|
| 57 |
+
BIWIRGBDID Testing_Still_010_010a_000021-a_171991229_rgb
|
| 58 |
+
BIWIRGBDID Testing_Still_010_010a_000054-a_171994564_rgb
|
| 59 |
+
BIWIRGBDID Testing_Still_010_010a_000063-a_171995649_rgb
|
| 60 |
+
BIWIRGBDID Testing_Still_010_010a_000072-a_171996480_rgb
|
| 61 |
+
BIWIRGBDID Testing_Still_010_010a_000085-a_171997897_rgb
|
| 62 |
+
BIWIRGBDID Testing_Still_010_010a_000087-a_171998064_rgb
|
| 63 |
+
BIWIRGBDID Testing_Still_010_010a_000098-a_171999229_rgb
|
| 64 |
+
BIWIRGBDID Testing_Still_010_010a_000107-a_172000148_rgb
|
| 65 |
+
BIWIRGBDID Testing_Still_010_010a_000119-a_172001396_rgb
|
| 66 |
+
BIWIRGBDID Testing_Still_011_011a_000023-a_200194778_rgb
|
| 67 |
+
BIWIRGBDID Testing_Still_011_011a_000088-a_200201941_rgb
|
| 68 |
+
BIWIRGBDID Testing_Still_011_011a_000101-a_200203442_rgb
|
| 69 |
+
BIWIRGBDID Testing_Still_011_011a_000104-a_200203778_rgb
|
| 70 |
+
BIWIRGBDID Testing_Still_011_011a_000116-a_200205277_rgb
|
| 71 |
+
BIWIRGBDID Testing_Still_016_016a_000000-a_188165678_rgb
|
| 72 |
+
BIWIRGBDID Testing_Still_016_016a_000026-a_188168426_rgb
|
| 73 |
+
BIWIRGBDID Testing_Still_016_016a_000043-a_188170258_rgb
|
| 74 |
+
BIWIRGBDID Testing_Still_016_016a_000052-a_188171178_rgb
|
| 75 |
+
BIWIRGBDID Testing_Still_016_016a_000114-a_188177758_rgb
|
| 76 |
+
BIWIRGBDID Testing_Still_017_017a_000004-a_191551441_rgb
|
| 77 |
+
BIWIRGBDID Testing_Still_017_017a_000050-a_191556773_rgb
|
| 78 |
+
BIWIRGBDID Testing_Still_017_017a_000055-a_191557272_rgb
|
| 79 |
+
BIWIRGBDID Testing_Still_017_017a_000059-a_191557688_rgb
|
| 80 |
+
BIWIRGBDID Testing_Still_017_017a_000073-a_191559189_rgb
|
| 81 |
+
BIWIRGBDID Testing_Still_017_017a_000099-a_191562440_rgb
|
| 82 |
+
BIWIRGBDID Testing_Still_017_017a_000102-a_191562688_rgb
|
| 83 |
+
BIWIRGBDID Testing_Still_017_017a_000103-a_191562772_rgb
|
| 84 |
+
BIWIRGBDID Testing_Still_017_017a_000122-a_191564772_rgb
|
| 85 |
+
BIWIRGBDID Testing_Still_019_019a_000024-a_107178865_rgb
|
| 86 |
+
BIWIRGBDID Testing_Still_019_019a_000081-a_107184780_rgb
|
| 87 |
+
BIWIRGBDID Testing_Still_019_019a_000086-a_107185364_rgb
|
| 88 |
+
BIWIRGBDID Testing_Still_021_021a_000028-a_109006730_rgb
|
| 89 |
+
BIWIRGBDID Testing_Still_021_021a_000081-a_109012398_rgb
|
| 90 |
+
BIWIRGBDID Testing_Still_021_021a_000084-a_109012646_rgb
|
| 91 |
+
BIWIRGBDID Testing_Still_021_021a_000100-a_109014314_rgb
|
| 92 |
+
BIWIRGBDID Testing_Still_021_021a_000117-a_109016062_rgb
|
| 93 |
+
BIWIRGBDID Testing_Still_021_021a_000137-a_109018146_rgb
|
| 94 |
+
BIWIRGBDID Testing_Still_022_022a_000052-a_1214872830_rgb
|
| 95 |
+
BIWIRGBDID Testing_Still_022_022a_000053-a_1214872914_rgb
|
| 96 |
+
BIWIRGBDID Testing_Still_022_022a_000072-a_1214875162_rgb
|
| 97 |
+
BIWIRGBDID Testing_Still_023_023a_000003-a_197683355_rgb
|
| 98 |
+
BIWIRGBDID Testing_Still_023_023a_000012-a_197684603_rgb
|
| 99 |
+
BIWIRGBDID Testing_Still_023_023a_000024-a_197686103_rgb
|
| 100 |
+
BIWIRGBDID Testing_Still_023_023a_000027-a_197686523_rgb
|
| 101 |
+
BIWIRGBDID Testing_Still_023_023a_000096-a_197695603_rgb
|
| 102 |
+
BIWIRGBDID Testing_Still_023_023a_000128-a_197699687_rgb
|
| 103 |
+
BIWIRGBDID Testing_Still_023_023a_000138-a_197700939_rgb
|
| 104 |
+
BIWIRGBDID Testing_Still_024_024a_000027-a_197310296_rgb
|
| 105 |
+
BIWIRGBDID Testing_Still_024_024a_000033-a_197311045_rgb
|
| 106 |
+
BIWIRGBDID Testing_Still_024_024a_000046-a_197312628_rgb
|
| 107 |
+
BIWIRGBDID Testing_Still_024_024a_000054-a_197313712_rgb
|
| 108 |
+
BIWIRGBDID Testing_Still_024_024a_000076-a_197316628_rgb
|
| 109 |
+
BIWIRGBDID Testing_Still_024_024a_000093-a_197319128_rgb
|
| 110 |
+
BIWIRGBDID Testing_Still_026_026a_000001-a_110999274_rgb
|
| 111 |
+
BIWIRGBDID Testing_Still_026_026a_000003-a_110999522_rgb
|
| 112 |
+
BIWIRGBDID Testing_Still_026_026a_000017-a_111001106_rgb
|
| 113 |
+
BIWIRGBDID Testing_Still_026_026a_000027-a_111002106_rgb
|
| 114 |
+
BIWIRGBDID Testing_Still_026_026a_000029-a_111002354_rgb
|
| 115 |
+
BIWIRGBDID Testing_Still_026_026a_000072-a_111007106_rgb
|
| 116 |
+
BIWIRGBDID Testing_Still_026_026a_000077-a_111007606_rgb
|
| 117 |
+
BIWIRGBDID Testing_Still_026_026a_000110-a_111011022_rgb
|
| 118 |
+
BIWIRGBDID Testing_Still_026_026a_000112-a_111011270_rgb
|
| 119 |
+
BIWIRGBDID Testing_Still_027_027a_000012-a_111337418_rgb
|
| 120 |
+
BIWIRGBDID Testing_Still_027_027a_000018-a_111338086_rgb
|
| 121 |
+
BIWIRGBDID Testing_Still_027_027a_000031-a_111339750_rgb
|
| 122 |
+
BIWIRGBDID Testing_Still_027_027a_000033-a_111339918_rgb
|
| 123 |
+
BIWIRGBDID Testing_Still_027_027a_000044-a_111341169_rgb
|
| 124 |
+
BIWIRGBDID Testing_Still_027_027a_000045-a_111341333_rgb
|
| 125 |
+
BIWIRGBDID Testing_Still_027_027a_000119-a_111349249_rgb
|
| 126 |
+
BIWIRGBDID Testing_Still_027_027a_000134-a_111350917_rgb
|
| 127 |
+
BIWIRGBDID Testing_Still_027_027a_000142-a_111351749_rgb
|
| 128 |
+
BIWIRGBDID Testing_Still_030_030a_000012-a_1233788639_rgb
|
| 129 |
+
BIWIRGBDID Testing_Still_030_030a_000043-a_1233791891_rgb
|
| 130 |
+
BIWIRGBDID Testing_Still_030_030a_000072-a_1233795471_rgb
|
| 131 |
+
BIWIRGBDID Testing_Still_030_030a_000091-a_1233797723_rgb
|
| 132 |
+
BIWIRGBDID Testing_Still_030_030a_000096-a_1233798307_rgb
|
| 133 |
+
BIWIRGBDID Testing_Still_030_030a_000107-a_1233799471_rgb
|
| 134 |
+
BIWIRGBDID Testing_Still_030_030a_000120-a_1233800971_rgb
|
| 135 |
+
BIWIRGBDID Testing_Still_030_030a_000127-a_1233801807_rgb
|
| 136 |
+
BIWIRGBDID Testing_Still_031_031a_000004-a_186779771_rgb
|
| 137 |
+
BIWIRGBDID Testing_Still_031_031a_000014-a_186780855_rgb
|
| 138 |
+
BIWIRGBDID Testing_Still_031_031a_000036-a_186783439_rgb
|
| 139 |
+
BIWIRGBDID Testing_Still_031_031a_000050-a_186784855_rgb
|
| 140 |
+
BIWIRGBDID Testing_Still_031_031a_000086-a_186788523_rgb
|
| 141 |
+
BIWIRGBDID Testing_Still_031_031a_000098-a_186789771_rgb
|
| 142 |
+
BIWIRGBDID Testing_Still_031_031a_000102-a_186790186_rgb
|
| 143 |
+
BIWIRGBDID Testing_Still_031_031a_000120-a_186791938_rgb
|
| 144 |
+
BIWIRGBDID Testing_Still_032_032a_000032-a_111072184_rgb
|
| 145 |
+
BIWIRGBDID Testing_Still_032_032a_000049-a_111074016_rgb
|
| 146 |
+
BIWIRGBDID Testing_Still_032_032a_000112-a_111080852_rgb
|
| 147 |
+
BIWIRGBDID Testing_Still_032_032a_000122-a_111081932_rgb
|
| 148 |
+
BIWIRGBDID Testing_Still_036_036a_000011-a_107768409_rgb
|
| 149 |
+
BIWIRGBDID Testing_Still_036_036a_000016-a_107768826_rgb
|
| 150 |
+
BIWIRGBDID Testing_Still_036_036a_000024-a_107769661_rgb
|
| 151 |
+
BIWIRGBDID Testing_Still_036_036a_000071-a_107774662_rgb
|
| 152 |
+
BIWIRGBDID Testing_Still_036_036a_000077-a_107775325_rgb
|
| 153 |
+
BIWIRGBDID Testing_Still_036_036a_000087-a_107776409_rgb
|
| 154 |
+
BIWIRGBDID Testing_Still_037_037a_000040-a_191434279_rgb
|
| 155 |
+
BIWIRGBDID Testing_Still_037_037a_000044-a_191434696_rgb
|
| 156 |
+
BIWIRGBDID Testing_Still_037_037a_000074-a_191438032_rgb
|
| 157 |
+
BIWIRGBDID Testing_Still_037_037a_000091-a_191439695_rgb
|
| 158 |
+
BIWIRGBDID Testing_Still_037_037a_000109-a_191441696_rgb
|
| 159 |
+
BIWIRGBDID Testing_Still_037_037a_000111-a_191442031_rgb
|
| 160 |
+
BIWIRGBDID Testing_Still_037_037a_000123-a_191443363_rgb
|
| 161 |
+
BIWIRGBDID Testing_Still_039_039a_000006-a_1219302217_rgb
|
| 162 |
+
BIWIRGBDID Testing_Still_039_039a_000015-a_1219303217_rgb
|
| 163 |
+
BIWIRGBDID Testing_Still_039_039a_000044-a_1219306217_rgb
|
| 164 |
+
BIWIRGBDID Testing_Still_039_039a_000058-a_1219307717_rgb
|
| 165 |
+
BIWIRGBDID Testing_Still_039_039a_000090-a_1219310969_rgb
|
| 166 |
+
BIWIRGBDID Testing_Still_039_039a_000097-a_1219311885_rgb
|
| 167 |
+
BIWIRGBDID Testing_Still_039_039a_000122-a_1219314717_rgb
|
| 168 |
+
BIWIRGBDID Testing_Still_039_039a_000144-a_1219317049_rgb
|
| 169 |
+
BIWIRGBDID Testing_Still_040_040a_000040-a_1231702678_rgb
|
| 170 |
+
BIWIRGBDID Testing_Still_040_040a_000046-a_1231703346_rgb
|
| 171 |
+
BIWIRGBDID Testing_Still_040_040a_000073-a_1231706514_rgb
|
| 172 |
+
BIWIRGBDID Testing_Still_040_040a_000102-a_1231710014_rgb
|
| 173 |
+
BIWIRGBDID Testing_Still_040_040a_000121-a_1231712093_rgb
|
| 174 |
+
BIWIRGBDID Testing_Still_040_040a_000122-a_1231712177_rgb
|
| 175 |
+
BIWIRGBDID Testing_Still_040_040a_000136-a_1231713846_rgb
|
| 176 |
+
BIWIRGBDID Testing_Still_044_044a_000002-a_1214165465_rgb
|
| 177 |
+
BIWIRGBDID Testing_Still_044_044a_000046-a_1214170130_rgb
|
| 178 |
+
BIWIRGBDID Testing_Still_044_044a_000059-a_1214171381_rgb
|
| 179 |
+
BIWIRGBDID Testing_Still_044_044a_000070-a_1214172549_rgb
|
| 180 |
+
BIWIRGBDID Testing_Still_044_044a_000089-a_1214174465_rgb
|
| 181 |
+
BIWIRGBDID Testing_Still_044_044a_000103-a_1214175881_rgb
|
| 182 |
+
BIWIRGBDID Testing_Still_046_046a_000006-a_111209841_rgb
|
| 183 |
+
BIWIRGBDID Testing_Still_046_046a_000021-a_111211425_rgb
|
| 184 |
+
BIWIRGBDID Testing_Still_046_046a_000040-a_111213425_rgb
|
| 185 |
+
BIWIRGBDID Testing_Still_046_046a_000071-a_111216841_rgb
|
| 186 |
+
BIWIRGBDID Testing_Still_046_046a_000078-a_111217509_rgb
|
| 187 |
+
BIWIRGBDID Testing_Still_046_046a_000079-a_111217676_rgb
|
| 188 |
+
BIWIRGBDID Testing_Still_047_047a_000067-a_1214094887_rgb
|
| 189 |
+
BIWIRGBDID Testing_Still_047_047a_000077-a_1214095887_rgb
|
| 190 |
+
BIWIRGBDID Testing_Still_047_047a_000090-a_1214097135_rgb
|
| 191 |
+
BIWIRGBDID Testing_Still_047_047a_000120-a_1214100468_rgb
|
| 192 |
+
BIWIRGBDID Testing_Still_049_049a_000000-a_201357193_rgb
|
| 193 |
+
BIWIRGBDID Testing_Still_049_049a_000020-a_201359362_rgb
|
| 194 |
+
BIWIRGBDID Testing_Still_049_049a_000026-a_201360029_rgb
|
| 195 |
+
BIWIRGBDID Testing_Still_049_049a_000028-a_201360193_rgb
|
| 196 |
+
BIWIRGBDID Testing_Still_049_049a_000069-a_201364613_rgb
|
| 197 |
+
BIWIRGBDID Testing_Still_049_049a_000073-a_201365029_rgb
|
| 198 |
+
BIWIRGBDID Testing_Still_049_049a_000083-a_201366113_rgb
|
| 199 |
+
BIWIRGBDID Testing_Still_049_049a_000088-a_201366613_rgb
|
| 200 |
+
BIWIRGBDID Testing_Walking_000_000b_000008-a_111676306_rgb
|
| 201 |
+
BIWIRGBDID Testing_Walking_000_000b_000032-a_111678306_rgb
|
| 202 |
+
BIWIRGBDID Testing_Walking_000_000b_000043-a_111679306_rgb
|
| 203 |
+
BIWIRGBDID Testing_Walking_000_000b_000064-a_111681390_rgb
|
| 204 |
+
BIWIRGBDID Testing_Walking_000_000b_000066-a_111681638_rgb
|
| 205 |
+
BIWIRGBDID Testing_Walking_000_000b_000085-a_111683806_rgb
|
| 206 |
+
BIWIRGBDID Testing_Walking_000_000b_000110-a_111686054_rgb
|
| 207 |
+
BIWIRGBDID Testing_Walking_000_000b_000139-a_111688470_rgb
|
| 208 |
+
BIWIRGBDID Testing_Walking_000_000b_000158-a_111690054_rgb
|
| 209 |
+
BIWIRGBDID Testing_Walking_000_000b_000159-a_111690138_rgb
|
| 210 |
+
BIWIRGBDID Testing_Walking_000_000b_000182-a_111692390_rgb
|
| 211 |
+
BIWIRGBDID Testing_Walking_000_000b_000192-a_111693390_rgb
|
| 212 |
+
BIWIRGBDID Testing_Walking_000_000b_000234-a_111697222_rgb
|
| 213 |
+
BIWIRGBDID Testing_Walking_000_000b_000258-a_111699221_rgb
|
| 214 |
+
BIWIRGBDID Testing_Walking_000_000b_000261-a_111699469_rgb
|
| 215 |
+
BIWIRGBDID Testing_Walking_000_000b_000265-a_111699886_rgb
|
| 216 |
+
BIWIRGBDID Testing_Walking_000_000b_000290-a_111702469_rgb
|
| 217 |
+
BIWIRGBDID Testing_Walking_000_000b_000291-a_111702553_rgb
|
| 218 |
+
BIWIRGBDID Testing_Walking_000_000b_000349-a_111708305_rgb
|
| 219 |
+
BIWIRGBDID Testing_Walking_000_000b_000355-a_111708805_rgb
|
| 220 |
+
BIWIRGBDID Testing_Walking_000_000b_000370-a_111710137_rgb
|
| 221 |
+
BIWIRGBDID Testing_Walking_000_000b_000372-a_111710305_rgb
|
| 222 |
+
BIWIRGBDID Testing_Walking_000_000b_000406-a_111713721_rgb
|
| 223 |
+
BIWIRGBDID Testing_Walking_000_000b_000410-a_111714137_rgb
|
| 224 |
+
BIWIRGBDID Testing_Walking_000_000b_000416-a_111714637_rgb
|
| 225 |
+
BIWIRGBDID Testing_Walking_000_000b_000420-a_111714969_rgb
|
| 226 |
+
BIWIRGBDID Testing_Walking_000_000b_000439-a_111716637_rgb
|
| 227 |
+
BIWIRGBDID Testing_Walking_001_001b_000002-a_106302670_rgb
|
| 228 |
+
BIWIRGBDID Testing_Walking_001_001b_000004-a_106302835_rgb
|
| 229 |
+
BIWIRGBDID Testing_Walking_001_001b_000011-a_106303418_rgb
|
| 230 |
+
BIWIRGBDID Testing_Walking_001_001b_000015-a_106303834_rgb
|
| 231 |
+
BIWIRGBDID Testing_Walking_001_001b_000030-a_106305254_rgb
|
| 232 |
+
BIWIRGBDID Testing_Walking_001_001b_000046-a_106306754_rgb
|
| 233 |
+
BIWIRGBDID Testing_Walking_001_001b_000083-a_106310586_rgb
|
| 234 |
+
BIWIRGBDID Testing_Walking_001_001b_000158-a_106317918_rgb
|
| 235 |
+
BIWIRGBDID Testing_Walking_001_001b_000177-a_106319586_rgb
|
| 236 |
+
BIWIRGBDID Testing_Walking_001_001b_000181-a_106319918_rgb
|
| 237 |
+
BIWIRGBDID Testing_Walking_001_001b_000205-a_106322170_rgb
|
| 238 |
+
BIWIRGBDID Testing_Walking_001_001b_000249-a_106327250_rgb
|
| 239 |
+
BIWIRGBDID Testing_Walking_001_001b_000289-a_106331834_rgb
|
| 240 |
+
BIWIRGBDID Testing_Walking_001_001b_000303-a_106333334_rgb
|
| 241 |
+
BIWIRGBDID Testing_Walking_001_001b_000329-a_106336502_rgb
|
| 242 |
+
BIWIRGBDID Testing_Walking_001_001b_000357-a_106339586_rgb
|
| 243 |
+
BIWIRGBDID Testing_Walking_001_001b_000360-a_106339917_rgb
|
| 244 |
+
BIWIRGBDID Testing_Walking_001_001b_000371-a_106341165_rgb
|
| 245 |
+
BIWIRGBDID Testing_Walking_002_002b_000004-a_1233001421_rgb
|
| 246 |
+
BIWIRGBDID Testing_Walking_002_002b_000018-a_1233002673_rgb
|
| 247 |
+
BIWIRGBDID Testing_Walking_002_002b_000024-a_1233003257_rgb
|
| 248 |
+
BIWIRGBDID Testing_Walking_002_002b_000032-a_1233004005_rgb
|
| 249 |
+
BIWIRGBDID Testing_Walking_002_002b_000059-a_1233006589_rgb
|
| 250 |
+
BIWIRGBDID Testing_Walking_002_002b_000069-a_1233007673_rgb
|
| 251 |
+
BIWIRGBDID Testing_Walking_002_002b_000130-a_1233014089_rgb
|
| 252 |
+
BIWIRGBDID Testing_Walking_002_002b_000159-a_1233017257_rgb
|
| 253 |
+
BIWIRGBDID Testing_Walking_002_002b_000169-a_1233018337_rgb
|
| 254 |
+
BIWIRGBDID Testing_Walking_002_002b_000170-a_1233018421_rgb
|
| 255 |
+
BIWIRGBDID Testing_Walking_002_002b_000188-a_1233020337_rgb
|
| 256 |
+
BIWIRGBDID Testing_Walking_002_002b_000220-a_1233024004_rgb
|
| 257 |
+
BIWIRGBDID Testing_Walking_002_002b_000222-a_1233024172_rgb
|
| 258 |
+
BIWIRGBDID Testing_Walking_002_002b_000232-a_1233025336_rgb
|
| 259 |
+
BIWIRGBDID Testing_Walking_002_002b_000287-a_1233031336_rgb
|
| 260 |
+
BIWIRGBDID Testing_Walking_002_002b_000303-a_1233033088_rgb
|
| 261 |
+
BIWIRGBDID Testing_Walking_002_002b_000308-a_1233033504_rgb
|
| 262 |
+
BIWIRGBDID Testing_Walking_002_002b_000318-a_1233034420_rgb
|
| 263 |
+
BIWIRGBDID Testing_Walking_002_002b_000338-a_1233036588_rgb
|
| 264 |
+
BIWIRGBDID Testing_Walking_002_002b_000340-a_1233036836_rgb
|
| 265 |
+
BIWIRGBDID Testing_Walking_002_002b_000355-a_1233038504_rgb
|
| 266 |
+
BIWIRGBDID Testing_Walking_003_003b_000025-a_621819949_rgb
|
| 267 |
+
BIWIRGBDID Testing_Walking_003_003b_000054-a_621823117_rgb
|
| 268 |
+
BIWIRGBDID Testing_Walking_003_003b_000112-a_621829533_rgb
|
| 269 |
+
BIWIRGBDID Testing_Walking_003_003b_000113-a_621829617_rgb
|
| 270 |
+
BIWIRGBDID Testing_Walking_003_003b_000173-a_621836365_rgb
|
| 271 |
+
BIWIRGBDID Testing_Walking_003_003b_000179-a_621836949_rgb
|
| 272 |
+
BIWIRGBDID Testing_Walking_003_003b_000189-a_621837781_rgb
|
| 273 |
+
BIWIRGBDID Testing_Walking_003_003b_000191-a_621837949_rgb
|
| 274 |
+
BIWIRGBDID Testing_Walking_003_003b_000225-a_621842201_rgb
|
| 275 |
+
BIWIRGBDID Testing_Walking_003_003b_000274-a_621847616_rgb
|
| 276 |
+
BIWIRGBDID Testing_Walking_003_003b_000275-a_621847701_rgb
|
| 277 |
+
BIWIRGBDID Testing_Walking_003_003b_000278-a_621847948_rgb
|
| 278 |
+
BIWIRGBDID Testing_Walking_003_003b_000306-a_621851116_rgb
|
| 279 |
+
BIWIRGBDID Testing_Walking_003_003b_000323-a_621852948_rgb
|
| 280 |
+
BIWIRGBDID Testing_Walking_003_003b_000335-a_621854032_rgb
|
| 281 |
+
BIWIRGBDID Testing_Walking_006_006b_000047-a_1224090658_rgb
|
| 282 |
+
BIWIRGBDID Testing_Walking_006_006b_000048-a_1224090742_rgb
|
| 283 |
+
BIWIRGBDID Testing_Walking_006_006b_000053-a_1224091325_rgb
|
| 284 |
+
BIWIRGBDID Testing_Walking_006_006b_000080-a_1224093993_rgb
|
| 285 |
+
BIWIRGBDID Testing_Walking_006_006b_000106-a_1224096741_rgb
|
| 286 |
+
BIWIRGBDID Testing_Walking_006_006b_000112-a_1224097325_rgb
|
| 287 |
+
BIWIRGBDID Testing_Walking_006_006b_000118-a_1224097825_rgb
|
| 288 |
+
BIWIRGBDID Testing_Walking_006_006b_000119-a_1224097909_rgb
|
| 289 |
+
BIWIRGBDID Testing_Walking_006_006b_000127-a_1224098657_rgb
|
| 290 |
+
BIWIRGBDID Testing_Walking_006_006b_000158-a_1224101325_rgb
|
| 291 |
+
BIWIRGBDID Testing_Walking_006_006b_000168-a_1224102241_rgb
|
| 292 |
+
BIWIRGBDID Testing_Walking_006_006b_000205-a_1224106073_rgb
|
| 293 |
+
BIWIRGBDID Testing_Walking_006_006b_000236-a_1224108825_rgb
|
| 294 |
+
BIWIRGBDID Testing_Walking_006_006b_000257-a_1224110573_rgb
|
| 295 |
+
BIWIRGBDID Testing_Walking_006_006b_000264-a_1224111157_rgb
|
| 296 |
+
BIWIRGBDID Testing_Walking_006_006b_000289-a_1224113825_rgb
|
| 297 |
+
BIWIRGBDID Testing_Walking_006_006b_000396-a_1224124073_rgb
|
| 298 |
+
BIWIRGBDID Testing_Walking_008_008b_000001-a_190329700_rgb
|
| 299 |
+
BIWIRGBDID Testing_Walking_008_008b_000030-a_190332200_rgb
|
| 300 |
+
BIWIRGBDID Testing_Walking_008_008b_000065-a_190336032_rgb
|
| 301 |
+
BIWIRGBDID Testing_Walking_008_008b_000087-a_190338532_rgb
|
| 302 |
+
BIWIRGBDID Testing_Walking_008_008b_000090-a_190338865_rgb
|
| 303 |
+
BIWIRGBDID Testing_Walking_008_008b_000093-a_190339117_rgb
|
| 304 |
+
BIWIRGBDID Testing_Walking_008_008b_000103-a_190339948_rgb
|
| 305 |
+
BIWIRGBDID Testing_Walking_008_008b_000124-a_190341700_rgb
|
| 306 |
+
BIWIRGBDID Testing_Walking_008_008b_000170-a_190346116_rgb
|
| 307 |
+
BIWIRGBDID Testing_Walking_008_008b_000218-a_190351280_rgb
|
| 308 |
+
BIWIRGBDID Testing_Walking_008_008b_000231-a_190352364_rgb
|
| 309 |
+
BIWIRGBDID Testing_Walking_008_008b_000257-a_190354780_rgb
|
| 310 |
+
BIWIRGBDID Testing_Walking_008_008b_000268-a_190355700_rgb
|
| 311 |
+
BIWIRGBDID Testing_Walking_008_008b_000307-a_190359696_rgb
|
| 312 |
+
BIWIRGBDID Testing_Walking_008_008b_000309-a_190359864_rgb
|
| 313 |
+
BIWIRGBDID Testing_Walking_008_008b_000317-a_190360531_rgb
|
| 314 |
+
BIWIRGBDID Testing_Walking_008_008b_000329-a_190361615_rgb
|
| 315 |
+
BIWIRGBDID Testing_Walking_008_008b_000363-a_190364948_rgb
|
| 316 |
+
BIWIRGBDID Testing_Walking_010_010b_000011-a_172014564_rgb
|
| 317 |
+
BIWIRGBDID Testing_Walking_010_010b_000027-a_172015896_rgb
|
| 318 |
+
BIWIRGBDID Testing_Walking_010_010b_000030-a_172016144_rgb
|
| 319 |
+
BIWIRGBDID Testing_Walking_010_010b_000077-a_172020812_rgb
|
| 320 |
+
BIWIRGBDID Testing_Walking_010_010b_000142-a_172026812_rgb
|
| 321 |
+
BIWIRGBDID Testing_Walking_010_010b_000167-a_172029395_rgb
|
| 322 |
+
BIWIRGBDID Testing_Walking_010_010b_000176-a_172030396_rgb
|
| 323 |
+
BIWIRGBDID Testing_Walking_010_010b_000187-a_172031560_rgb
|
| 324 |
+
BIWIRGBDID Testing_Walking_010_010b_000212-a_172033980_rgb
|
| 325 |
+
BIWIRGBDID Testing_Walking_010_010b_000237-a_172036060_rgb
|
| 326 |
+
BIWIRGBDID Testing_Walking_010_010b_000255-a_172037727_rgb
|
| 327 |
+
BIWIRGBDID Testing_Walking_010_010b_000282-a_172040395_rgb
|
| 328 |
+
BIWIRGBDID Testing_Walking_010_010b_000321-a_172044395_rgb
|
| 329 |
+
BIWIRGBDID Testing_Walking_010_010b_000324-a_172044643_rgb
|
| 330 |
+
BIWIRGBDID Testing_Walking_010_010b_000330-a_172045311_rgb
|
| 331 |
+
BIWIRGBDID Testing_Walking_010_010b_000351-a_172047143_rgb
|
| 332 |
+
BIWIRGBDID Testing_Walking_010_010b_000352-a_172047227_rgb
|
| 333 |
+
BIWIRGBDID Testing_Walking_010_010b_000380-a_172049643_rgb
|
| 334 |
+
BIWIRGBDID Testing_Walking_010_010b_000382-a_172049811_rgb
|
| 335 |
+
BIWIRGBDID Testing_Walking_011_011b_000025-a_200221193_rgb
|
| 336 |
+
BIWIRGBDID Testing_Walking_011_011b_000033-a_200222025_rgb
|
| 337 |
+
BIWIRGBDID Testing_Walking_011_011b_000044-a_200223193_rgb
|
| 338 |
+
BIWIRGBDID Testing_Walking_011_011b_000049-a_200223693_rgb
|
| 339 |
+
BIWIRGBDID Testing_Walking_011_011b_000083-a_200227525_rgb
|
| 340 |
+
BIWIRGBDID Testing_Walking_011_011b_000086-a_200227857_rgb
|
| 341 |
+
BIWIRGBDID Testing_Walking_011_011b_000106-a_200230193_rgb
|
| 342 |
+
BIWIRGBDID Testing_Walking_011_011b_000125-a_200232025_rgb
|
| 343 |
+
BIWIRGBDID Testing_Walking_011_011b_000142-a_200233609_rgb
|
| 344 |
+
BIWIRGBDID Testing_Walking_011_011b_000175-a_200237109_rgb
|
| 345 |
+
BIWIRGBDID Testing_Walking_011_011b_000203-a_200240357_rgb
|
| 346 |
+
BIWIRGBDID Testing_Walking_011_011b_000204-a_200240441_rgb
|
| 347 |
+
BIWIRGBDID Testing_Walking_011_011b_000259-a_200246188_rgb
|
| 348 |
+
BIWIRGBDID Testing_Walking_011_011b_000266-a_200246857_rgb
|
| 349 |
+
BIWIRGBDID Testing_Walking_011_011b_000293-a_200249524_rgb
|
| 350 |
+
BIWIRGBDID Testing_Walking_011_011b_000301-a_200250440_rgb
|
| 351 |
+
BIWIRGBDID Testing_Walking_011_011b_000309-a_200251441_rgb
|
| 352 |
+
BIWIRGBDID Testing_Walking_011_011b_000321-a_200252772_rgb
|
| 353 |
+
BIWIRGBDID Testing_Walking_011_011b_000367-a_200257440_rgb
|
| 354 |
+
BIWIRGBDID Testing_Walking_011_011b_000393-a_200259772_rgb
|
| 355 |
+
BIWIRGBDID Testing_Walking_011_011b_000398-a_200260272_rgb
|
| 356 |
+
BIWIRGBDID Testing_Walking_011_011b_000430-a_200263772_rgb
|
| 357 |
+
BIWIRGBDID Testing_Walking_016_016b_000003-a_188187593_rgb
|
| 358 |
+
BIWIRGBDID Testing_Walking_016_016b_000004-a_188187673_rgb
|
| 359 |
+
BIWIRGBDID Testing_Walking_016_016b_000016-a_188188841_rgb
|
| 360 |
+
BIWIRGBDID Testing_Walking_016_016b_000040-a_188191173_rgb
|
| 361 |
+
BIWIRGBDID Testing_Walking_016_016b_000064-a_188193841_rgb
|
| 362 |
+
BIWIRGBDID Testing_Walking_016_016b_000082-a_188195673_rgb
|
| 363 |
+
BIWIRGBDID Testing_Walking_016_016b_000087-a_188196093_rgb
|
| 364 |
+
BIWIRGBDID Testing_Walking_016_016b_000108-a_188198509_rgb
|
| 365 |
+
BIWIRGBDID Testing_Walking_016_016b_000111-a_188198841_rgb
|
| 366 |
+
BIWIRGBDID Testing_Walking_016_016b_000141-a_188202009_rgb
|
| 367 |
+
BIWIRGBDID Testing_Walking_016_016b_000142-a_188202173_rgb
|
| 368 |
+
BIWIRGBDID Testing_Walking_016_016b_000161-a_188204173_rgb
|
| 369 |
+
BIWIRGBDID Testing_Walking_016_016b_000202-a_188208341_rgb
|
| 370 |
+
BIWIRGBDID Testing_Walking_016_016b_000206-a_188208841_rgb
|
| 371 |
+
BIWIRGBDID Testing_Walking_016_016b_000211-a_188209425_rgb
|
| 372 |
+
BIWIRGBDID Testing_Walking_016_016b_000216-a_188209841_rgb
|
| 373 |
+
BIWIRGBDID Testing_Walking_016_016b_000235-a_188211757_rgb
|
| 374 |
+
BIWIRGBDID Testing_Walking_016_016b_000238-a_188212173_rgb
|
| 375 |
+
BIWIRGBDID Testing_Walking_016_016b_000253-a_188213757_rgb
|
| 376 |
+
BIWIRGBDID Testing_Walking_016_016b_000268-a_188215341_rgb
|
| 377 |
+
BIWIRGBDID Testing_Walking_016_016b_000287-a_188217508_rgb
|
| 378 |
+
BIWIRGBDID Testing_Walking_016_016b_000298-a_188218672_rgb
|
| 379 |
+
BIWIRGBDID Testing_Walking_016_016b_000308-a_188219841_rgb
|
| 380 |
+
BIWIRGBDID Testing_Walking_016_016b_000310-a_188220004_rgb
|
| 381 |
+
BIWIRGBDID Testing_Walking_016_016b_000313-a_188220341_rgb
|
| 382 |
+
BIWIRGBDID Testing_Walking_016_016b_000318-a_188220925_rgb
|
| 383 |
+
BIWIRGBDID Testing_Walking_016_016b_000347-a_188224005_rgb
|
| 384 |
+
BIWIRGBDID Testing_Walking_017_017b_000008-a_191574104_rgb
|
| 385 |
+
BIWIRGBDID Testing_Walking_017_017b_000011-a_191574356_rgb
|
| 386 |
+
BIWIRGBDID Testing_Walking_017_017b_000039-a_191577604_rgb
|
| 387 |
+
BIWIRGBDID Testing_Walking_017_017b_000090-a_191583104_rgb
|
| 388 |
+
BIWIRGBDID Testing_Walking_017_017b_000093-a_191583436_rgb
|
| 389 |
+
BIWIRGBDID Testing_Walking_017_017b_000095-a_191583937_rgb
|
| 390 |
+
BIWIRGBDID Testing_Walking_017_017b_000100-a_191584436_rgb
|
| 391 |
+
BIWIRGBDID Testing_Walking_017_017b_000119-a_191586272_rgb
|
| 392 |
+
BIWIRGBDID Testing_Walking_017_017b_000179-a_191592103_rgb
|
| 393 |
+
BIWIRGBDID Testing_Walking_017_017b_000181-a_191592272_rgb
|
| 394 |
+
BIWIRGBDID Testing_Walking_017_017b_000201-a_191594436_rgb
|
| 395 |
+
BIWIRGBDID Testing_Walking_017_017b_000227-a_191597355_rgb
|
| 396 |
+
BIWIRGBDID Testing_Walking_017_017b_000241-a_191598851_rgb
|
| 397 |
+
BIWIRGBDID Testing_Walking_017_017b_000242-a_191598935_rgb
|
| 398 |
+
BIWIRGBDID Testing_Walking_017_017b_000252-a_191599935_rgb
|
| 399 |
+
BIWIRGBDID Testing_Walking_017_017b_000263-a_191600935_rgb
|
| 400 |
+
BIWIRGBDID Testing_Walking_017_017b_000275-a_191602187_rgb
|
| 401 |
+
BIWIRGBDID Testing_Walking_017_017b_000291-a_191603852_rgb
|
| 402 |
+
BIWIRGBDID Testing_Walking_017_017b_000306-a_191605435_rgb
|
| 403 |
+
BIWIRGBDID Testing_Walking_017_017b_000313-a_191606103_rgb
|
| 404 |
+
BIWIRGBDID Testing_Walking_017_017b_000317-a_191606436_rgb
|
| 405 |
+
BIWIRGBDID Testing_Walking_017_017b_000385-a_191613687_rgb
|
| 406 |
+
BIWIRGBDID Testing_Walking_017_017b_000422-a_191617351_rgb
|
| 407 |
+
BIWIRGBDID Testing_Walking_017_017b_000425-a_191617603_rgb
|
| 408 |
+
BIWIRGBDID Testing_Walking_017_017b_000462-a_191622103_rgb
|
| 409 |
+
BIWIRGBDID Testing_Walking_017_017b_000472-a_191623103_rgb
|
| 410 |
+
BIWIRGBDID Testing_Walking_017_017b_000496-a_191625603_rgb
|
| 411 |
+
BIWIRGBDID Testing_Walking_019_019b_000023-a_107227611_rgb
|
| 412 |
+
BIWIRGBDID Testing_Walking_019_019b_000038-a_107228863_rgb
|
| 413 |
+
BIWIRGBDID Testing_Walking_019_019b_000046-a_107229611_rgb
|
| 414 |
+
BIWIRGBDID Testing_Walking_019_019b_000047-a_107229695_rgb
|
| 415 |
+
BIWIRGBDID Testing_Walking_019_019b_000103-a_107235443_rgb
|
| 416 |
+
BIWIRGBDID Testing_Walking_019_019b_000128-a_107238027_rgb
|
| 417 |
+
BIWIRGBDID Testing_Walking_019_019b_000150-a_107240027_rgb
|
| 418 |
+
BIWIRGBDID Testing_Walking_019_019b_000183-a_107243027_rgb
|
| 419 |
+
BIWIRGBDID Testing_Walking_019_019b_000188-a_107243527_rgb
|
| 420 |
+
BIWIRGBDID Testing_Walking_019_019b_000210-a_107245695_rgb
|
| 421 |
+
BIWIRGBDID Testing_Walking_019_019b_000218-a_107246527_rgb
|
| 422 |
+
BIWIRGBDID Testing_Walking_019_019b_000254-a_107250359_rgb
|
| 423 |
+
BIWIRGBDID Testing_Walking_019_019b_000256-a_107250526_rgb
|
| 424 |
+
BIWIRGBDID Testing_Walking_019_019b_000269-a_107251694_rgb
|
| 425 |
+
BIWIRGBDID Testing_Walking_019_019b_000285-a_107253194_rgb
|
| 426 |
+
BIWIRGBDID Testing_Walking_019_019b_000310-a_107255358_rgb
|
| 427 |
+
BIWIRGBDID Testing_Walking_019_019b_000311-a_107255442_rgb
|
| 428 |
+
BIWIRGBDID Testing_Walking_019_019b_000313-a_107255610_rgb
|
| 429 |
+
BIWIRGBDID Testing_Walking_019_019b_000336-a_107258026_rgb
|
| 430 |
+
BIWIRGBDID Testing_Walking_019_019b_000347-a_107259194_rgb
|
| 431 |
+
BIWIRGBDID Testing_Walking_019_019b_000373-a_107261942_rgb
|
| 432 |
+
BIWIRGBDID Testing_Walking_019_019b_000410-a_107265526_rgb
|
| 433 |
+
BIWIRGBDID Testing_Walking_019_019b_000416-a_107266026_rgb
|
| 434 |
+
BIWIRGBDID Testing_Walking_019_019b_000433-a_107267526_rgb
|
| 435 |
+
BIWIRGBDID Testing_Walking_021_021b_000075-a_109053893_rgb
|
| 436 |
+
BIWIRGBDID Testing_Walking_021_021b_000089-a_109055313_rgb
|
| 437 |
+
BIWIRGBDID Testing_Walking_021_021b_000095-a_109055978_rgb
|
| 438 |
+
BIWIRGBDID Testing_Walking_021_021b_000105-a_109056977_rgb
|
| 439 |
+
BIWIRGBDID Testing_Walking_021_021b_000111-a_109057561_rgb
|
| 440 |
+
BIWIRGBDID Testing_Walking_021_021b_000144-a_109060813_rgb
|
| 441 |
+
BIWIRGBDID Testing_Walking_021_021b_000176-a_109063645_rgb
|
| 442 |
+
BIWIRGBDID Testing_Walking_021_021b_000235-a_109069309_rgb
|
| 443 |
+
BIWIRGBDID Testing_Walking_021_021b_000265-a_109072561_rgb
|
| 444 |
+
BIWIRGBDID Testing_Walking_021_021b_000305-a_109076060_rgb
|
| 445 |
+
BIWIRGBDID Testing_Walking_021_021b_000328-a_109078145_rgb
|
| 446 |
+
BIWIRGBDID Testing_Walking_021_021b_000356-a_109080976_rgb
|
| 447 |
+
BIWIRGBDID Testing_Walking_021_021b_000370-a_109082476_rgb
|
| 448 |
+
BIWIRGBDID Testing_Walking_021_021b_000392-a_109084724_rgb
|
| 449 |
+
BIWIRGBDID Testing_Walking_021_021b_000420-a_109087392_rgb
|
| 450 |
+
BIWIRGBDID Testing_Walking_021_021b_000440-a_109089060_rgb
|
| 451 |
+
BIWIRGBDID Testing_Walking_021_021b_000493-a_109093476_rgb
|
| 452 |
+
BIWIRGBDID Testing_Walking_022_022b_000032-a_1214892493_rgb
|
| 453 |
+
BIWIRGBDID Testing_Walking_022_022b_000043-a_1214893409_rgb
|
| 454 |
+
BIWIRGBDID Testing_Walking_022_022b_000052-a_1214894161_rgb
|
| 455 |
+
BIWIRGBDID Testing_Walking_022_022b_000054-a_1214894329_rgb
|
| 456 |
+
BIWIRGBDID Testing_Walking_022_022b_000058-a_1214894661_rgb
|
| 457 |
+
BIWIRGBDID Testing_Walking_022_022b_000087-a_1214897409_rgb
|
| 458 |
+
BIWIRGBDID Testing_Walking_022_022b_000088-a_1214897493_rgb
|
| 459 |
+
BIWIRGBDID Testing_Walking_022_022b_000107-a_1214899409_rgb
|
| 460 |
+
BIWIRGBDID Testing_Walking_022_022b_000126-a_1214901329_rgb
|
| 461 |
+
BIWIRGBDID Testing_Walking_022_022b_000147-a_1214903161_rgb
|
| 462 |
+
BIWIRGBDID Testing_Walking_022_022b_000183-a_1214906493_rgb
|
| 463 |
+
BIWIRGBDID Testing_Walking_022_022b_000185-a_1214906661_rgb
|
| 464 |
+
BIWIRGBDID Testing_Walking_022_022b_000200-a_1214908161_rgb
|
| 465 |
+
BIWIRGBDID Testing_Walking_022_022b_000201-a_1214908245_rgb
|
| 466 |
+
BIWIRGBDID Testing_Walking_022_022b_000206-a_1214908745_rgb
|
| 467 |
+
BIWIRGBDID Testing_Walking_022_022b_000256-a_1214913825_rgb
|
| 468 |
+
BIWIRGBDID Testing_Walking_022_022b_000279-a_1214915909_rgb
|
| 469 |
+
BIWIRGBDID Testing_Walking_022_022b_000288-a_1214916909_rgb
|
| 470 |
+
BIWIRGBDID Testing_Walking_022_022b_000320-a_1214920077_rgb
|
| 471 |
+
BIWIRGBDID Testing_Walking_022_022b_000344-a_1214922660_rgb
|
| 472 |
+
BIWIRGBDID Testing_Walking_022_022b_000360-a_1214924741_rgb
|
| 473 |
+
BIWIRGBDID Testing_Walking_022_022b_000368-a_1214925492_rgb
|
| 474 |
+
BIWIRGBDID Testing_Walking_022_022b_000375-a_1214926076_rgb
|
| 475 |
+
BIWIRGBDID Testing_Walking_022_022b_000379-a_1214926408_rgb
|
| 476 |
+
BIWIRGBDID Testing_Walking_022_022b_000404-a_1214928909_rgb
|
| 477 |
+
BIWIRGBDID Testing_Walking_022_022b_000405-a_1214928992_rgb
|
| 478 |
+
BIWIRGBDID Testing_Walking_022_022b_000437-a_1214931825_rgb
|
| 479 |
+
BIWIRGBDID Testing_Walking_022_022b_000452-a_1214933324_rgb
|
| 480 |
+
BIWIRGBDID Testing_Walking_022_022b_000459-a_1214934076_rgb
|
| 481 |
+
BIWIRGBDID Testing_Walking_023_023b_000009-a_197707271_rgb
|
| 482 |
+
BIWIRGBDID Testing_Walking_023_023b_000020-a_197708439_rgb
|
| 483 |
+
BIWIRGBDID Testing_Walking_023_023b_000029-a_197709270_rgb
|
| 484 |
+
BIWIRGBDID Testing_Walking_023_023b_000051-a_197711186_rgb
|
| 485 |
+
BIWIRGBDID Testing_Walking_023_023b_000063-a_197712438_rgb
|
| 486 |
+
BIWIRGBDID Testing_Walking_023_023b_000077-a_197714271_rgb
|
| 487 |
+
BIWIRGBDID Testing_Walking_023_023b_000085-a_197715354_rgb
|
| 488 |
+
BIWIRGBDID Testing_Walking_023_023b_000099-a_197717186_rgb
|
| 489 |
+
BIWIRGBDID Testing_Walking_023_023b_000112-a_197718854_rgb
|
| 490 |
+
BIWIRGBDID Testing_Walking_023_023b_000118-a_197719518_rgb
|
| 491 |
+
BIWIRGBDID Testing_Walking_023_023b_000119-a_197719602_rgb
|
| 492 |
+
BIWIRGBDID Testing_Walking_023_023b_000124-a_197720186_rgb
|
| 493 |
+
BIWIRGBDID Testing_Walking_023_023b_000130-a_197720686_rgb
|
| 494 |
+
BIWIRGBDID Testing_Walking_023_023b_000155-a_197722935_rgb
|
| 495 |
+
BIWIRGBDID Testing_Walking_023_023b_000181-a_197725602_rgb
|
| 496 |
+
BIWIRGBDID Testing_Walking_023_023b_000197-a_197727855_rgb
|
| 497 |
+
BIWIRGBDID Testing_Walking_023_023b_000211-a_197729770_rgb
|
| 498 |
+
BIWIRGBDID Testing_Walking_023_023b_000222-a_197731018_rgb
|
| 499 |
+
BIWIRGBDID Testing_Walking_023_023b_000233-a_197732186_rgb
|
| 500 |
+
BIWIRGBDID Testing_Walking_023_023b_000253-a_197733850_rgb
|
| 501 |
+
BIWIRGBDID Testing_Walking_023_023b_000261-a_197734602_rgb
|
| 502 |
+
BIWIRGBDID Testing_Walking_023_023b_000266-a_197735018_rgb
|
| 503 |
+
BIWIRGBDID Testing_Walking_023_023b_000276-a_197735934_rgb
|
| 504 |
+
BIWIRGBDID Testing_Walking_023_023b_000279-a_197736185_rgb
|
| 505 |
+
BIWIRGBDID Testing_Walking_023_023b_000285-a_197736686_rgb
|
| 506 |
+
BIWIRGBDID Testing_Walking_023_023b_000307-a_197738850_rgb
|
| 507 |
+
BIWIRGBDID Testing_Walking_023_023b_000310-a_197739186_rgb
|
| 508 |
+
BIWIRGBDID Testing_Walking_023_023b_000328-a_197741518_rgb
|
| 509 |
+
BIWIRGBDID Testing_Walking_023_023b_000356-a_197745433_rgb
|
| 510 |
+
BIWIRGBDID Testing_Walking_023_023b_000371-a_197747017_rgb
|
| 511 |
+
BIWIRGBDID Testing_Walking_023_023b_000379-a_197747685_rgb
|
| 512 |
+
BIWIRGBDID Testing_Walking_023_023b_000407-a_197750265_rgb
|
| 513 |
+
BIWIRGBDID Testing_Walking_023_023b_000416-a_197751018_rgb
|
| 514 |
+
BIWIRGBDID Testing_Walking_023_023b_000424-a_197751765_rgb
|
| 515 |
+
BIWIRGBDID Testing_Walking_023_023b_000438-a_197753517_rgb
|
| 516 |
+
BIWIRGBDID Testing_Walking_024_024b_000028-a_197337876_rgb
|
| 517 |
+
BIWIRGBDID Testing_Walking_024_024b_000035-a_197338544_rgb
|
| 518 |
+
BIWIRGBDID Testing_Walking_024_024b_000058-a_197341708_rgb
|
| 519 |
+
BIWIRGBDID Testing_Walking_024_024b_000061-a_197342128_rgb
|
| 520 |
+
BIWIRGBDID Testing_Walking_024_024b_000065-a_197342627_rgb
|
| 521 |
+
BIWIRGBDID Testing_Walking_024_024b_000074-a_197343628_rgb
|
| 522 |
+
BIWIRGBDID Testing_Walking_024_024b_000090-a_197345628_rgb
|
| 523 |
+
BIWIRGBDID Testing_Walking_024_024b_000132-a_197349628_rgb
|
| 524 |
+
BIWIRGBDID Testing_Walking_024_024b_000184-a_197356127_rgb
|
| 525 |
+
BIWIRGBDID Testing_Walking_024_024b_000210-a_197358459_rgb
|
| 526 |
+
BIWIRGBDID Testing_Walking_024_024b_000239-a_197361544_rgb
|
| 527 |
+
BIWIRGBDID Testing_Walking_024_024b_000259-a_197364043_rgb
|
| 528 |
+
BIWIRGBDID Testing_Walking_024_024b_000304-a_197369043_rgb
|
| 529 |
+
BIWIRGBDID Testing_Walking_024_024b_000355-a_197374707_rgb
|
| 530 |
+
BIWIRGBDID Testing_Walking_026_026b_000007-a_111022853_rgb
|
| 531 |
+
BIWIRGBDID Testing_Walking_026_026b_000025-a_111024354_rgb
|
| 532 |
+
BIWIRGBDID Testing_Walking_026_026b_000038-a_111025770_rgb
|
| 533 |
+
BIWIRGBDID Testing_Walking_026_026b_000061-a_111028185_rgb
|
| 534 |
+
BIWIRGBDID Testing_Walking_026_026b_000062-a_111028269_rgb
|
| 535 |
+
BIWIRGBDID Testing_Walking_026_026b_000067-a_111028769_rgb
|
| 536 |
+
BIWIRGBDID Testing_Walking_026_026b_000072-a_111029353_rgb
|
| 537 |
+
BIWIRGBDID Testing_Walking_026_026b_000083-a_111030521_rgb
|
| 538 |
+
BIWIRGBDID Testing_Walking_026_026b_000097-a_111031769_rgb
|
| 539 |
+
BIWIRGBDID Testing_Walking_026_026b_000102-a_111032186_rgb
|
| 540 |
+
BIWIRGBDID Testing_Walking_026_026b_000108-a_111032854_rgb
|
| 541 |
+
BIWIRGBDID Testing_Walking_026_026b_000114-a_111033354_rgb
|
| 542 |
+
BIWIRGBDID Testing_Walking_026_026b_000136-a_111035605_rgb
|
| 543 |
+
BIWIRGBDID Testing_Walking_026_026b_000160-a_111038185_rgb
|
| 544 |
+
BIWIRGBDID Testing_Walking_026_026b_000162-a_111038353_rgb
|
| 545 |
+
BIWIRGBDID Testing_Walking_026_026b_000179-a_111040105_rgb
|
| 546 |
+
BIWIRGBDID Testing_Walking_026_026b_000213-a_111043769_rgb
|
| 547 |
+
BIWIRGBDID Testing_Walking_026_026b_000215-a_111044021_rgb
|
| 548 |
+
BIWIRGBDID Testing_Walking_026_026b_000221-a_111044521_rgb
|
| 549 |
+
BIWIRGBDID Testing_Walking_026_026b_000232-a_111045602_rgb
|
| 550 |
+
BIWIRGBDID Testing_Walking_026_026b_000247-a_111047185_rgb
|
| 551 |
+
BIWIRGBDID Testing_Walking_026_026b_000249-a_111047353_rgb
|
| 552 |
+
BIWIRGBDID Testing_Walking_026_026b_000265-a_111049021_rgb
|
| 553 |
+
BIWIRGBDID Testing_Walking_026_026b_000275-a_111050101_rgb
|
| 554 |
+
BIWIRGBDID Testing_Walking_026_026b_000333-a_111055601_rgb
|
| 555 |
+
BIWIRGBDID Testing_Walking_027_027b_000001-a_111418663_rgb
|
| 556 |
+
BIWIRGBDID Testing_Walking_027_027b_000017-a_111419995_rgb
|
| 557 |
+
BIWIRGBDID Testing_Walking_027_027b_000028-a_111420911_rgb
|
| 558 |
+
BIWIRGBDID Testing_Walking_027_027b_000049-a_111422912_rgb
|
| 559 |
+
BIWIRGBDID Testing_Walking_027_027b_000062-a_111424247_rgb
|
| 560 |
+
BIWIRGBDID Testing_Walking_027_027b_000072-a_111425327_rgb
|
| 561 |
+
BIWIRGBDID Testing_Walking_027_027b_000080-a_111426163_rgb
|
| 562 |
+
BIWIRGBDID Testing_Walking_027_027b_000088-a_111426995_rgb
|
| 563 |
+
BIWIRGBDID Testing_Walking_027_027b_000098-a_111428079_rgb
|
| 564 |
+
BIWIRGBDID Testing_Walking_027_027b_000132-a_111431163_rgb
|
| 565 |
+
BIWIRGBDID Testing_Walking_027_027b_000198-a_111436827_rgb
|
| 566 |
+
BIWIRGBDID Testing_Walking_027_027b_000203-a_111437412_rgb
|
| 567 |
+
BIWIRGBDID Testing_Walking_027_027b_000205-a_111437579_rgb
|
| 568 |
+
BIWIRGBDID Testing_Walking_027_027b_000207-a_111437827_rgb
|
| 569 |
+
BIWIRGBDID Testing_Walking_027_027b_000218-a_111438911_rgb
|
| 570 |
+
BIWIRGBDID Testing_Walking_027_027b_000240-a_111441411_rgb
|
| 571 |
+
BIWIRGBDID Testing_Walking_027_027b_000313-a_111448663_rgb
|
| 572 |
+
BIWIRGBDID Testing_Walking_027_027b_000323-a_111449663_rgb
|
| 573 |
+
BIWIRGBDID Testing_Walking_027_027b_000338-a_111451243_rgb
|
| 574 |
+
BIWIRGBDID Testing_Walking_027_027b_000350-a_111452495_rgb
|
| 575 |
+
BIWIRGBDID Testing_Walking_027_027b_000373-a_111454911_rgb
|
| 576 |
+
BIWIRGBDID Testing_Walking_027_027b_000380-a_111455495_rgb
|
| 577 |
+
BIWIRGBDID Testing_Walking_027_027b_000382-a_111455659_rgb
|
| 578 |
+
BIWIRGBDID Testing_Walking_027_027b_000405-a_111457743_rgb
|
| 579 |
+
BIWIRGBDID Testing_Walking_027_027b_000412-a_111458495_rgb
|
| 580 |
+
BIWIRGBDID Testing_Walking_030_030b_000015-a_1233809555_rgb
|
| 581 |
+
BIWIRGBDID Testing_Walking_030_030b_000022-a_1233810139_rgb
|
| 582 |
+
BIWIRGBDID Testing_Walking_030_030b_000068-a_1233815139_rgb
|
| 583 |
+
BIWIRGBDID Testing_Walking_030_030b_000075-a_1233815970_rgb
|
| 584 |
+
BIWIRGBDID Testing_Walking_030_030b_000077-a_1233816139_rgb
|
| 585 |
+
BIWIRGBDID Testing_Walking_030_030b_000079-a_1233816307_rgb
|
| 586 |
+
BIWIRGBDID Testing_Walking_030_030b_000102-a_1233818638_rgb
|
| 587 |
+
BIWIRGBDID Testing_Walking_030_030b_000129-a_1233820886_rgb
|
| 588 |
+
BIWIRGBDID Testing_Walking_030_030b_000156-a_1233823886_rgb
|
| 589 |
+
BIWIRGBDID Testing_Walking_030_030b_000168-a_1233825054_rgb
|
| 590 |
+
BIWIRGBDID Testing_Walking_030_030b_000182-a_1233826638_rgb
|
| 591 |
+
BIWIRGBDID Testing_Walking_030_030b_000197-a_1233828222_rgb
|
| 592 |
+
BIWIRGBDID Testing_Walking_030_030b_000216-a_1233830138_rgb
|
| 593 |
+
BIWIRGBDID Testing_Walking_030_030b_000219-a_1233830386_rgb
|
| 594 |
+
BIWIRGBDID Testing_Walking_030_030b_000223-a_1233830722_rgb
|
| 595 |
+
BIWIRGBDID Testing_Walking_030_030b_000226-a_1233830970_rgb
|
| 596 |
+
BIWIRGBDID Testing_Walking_030_030b_000234-a_1233831722_rgb
|
| 597 |
+
BIWIRGBDID Testing_Walking_030_030b_000248-a_1233832886_rgb
|
| 598 |
+
BIWIRGBDID Testing_Walking_030_030b_000299-a_1233838054_rgb
|
| 599 |
+
BIWIRGBDID Testing_Walking_030_030b_000334-a_1233841722_rgb
|
| 600 |
+
BIWIRGBDID Testing_Walking_030_030b_000344-a_1233842554_rgb
|
| 601 |
+
BIWIRGBDID Testing_Walking_030_030b_000350-a_1233843218_rgb
|
| 602 |
+
BIWIRGBDID Testing_Walking_030_030b_000351-a_1233843386_rgb
|
| 603 |
+
BIWIRGBDID Testing_Walking_030_030b_000357-a_1233844302_rgb
|
| 604 |
+
BIWIRGBDID Testing_Walking_030_030b_000367-a_1233845302_rgb
|
| 605 |
+
BIWIRGBDID Testing_Walking_030_030b_000398-a_1233848638_rgb
|
| 606 |
+
BIWIRGBDID Testing_Walking_031_031b_000001-a_186803438_rgb
|
| 607 |
+
BIWIRGBDID Testing_Walking_031_031b_000005-a_186803770_rgb
|
| 608 |
+
BIWIRGBDID Testing_Walking_031_031b_000054-a_186808770_rgb
|
| 609 |
+
BIWIRGBDID Testing_Walking_031_031b_000069-a_186810354_rgb
|
| 610 |
+
BIWIRGBDID Testing_Walking_031_031b_000079-a_186811438_rgb
|
| 611 |
+
BIWIRGBDID Testing_Walking_031_031b_000125-a_186816438_rgb
|
| 612 |
+
BIWIRGBDID Testing_Walking_031_031b_000132-a_186817102_rgb
|
| 613 |
+
BIWIRGBDID Testing_Walking_031_031b_000165-a_186820434_rgb
|
| 614 |
+
BIWIRGBDID Testing_Walking_031_031b_000176-a_186821518_rgb
|
| 615 |
+
BIWIRGBDID Testing_Walking_031_031b_000197-a_186823770_rgb
|
| 616 |
+
BIWIRGBDID Testing_Walking_031_031b_000215-a_186825601_rgb
|
| 617 |
+
BIWIRGBDID Testing_Walking_031_031b_000221-a_186826102_rgb
|
| 618 |
+
BIWIRGBDID Testing_Walking_031_031b_000245-a_186828602_rgb
|
| 619 |
+
BIWIRGBDID Testing_Walking_031_031b_000259-a_186829934_rgb
|
| 620 |
+
BIWIRGBDID Testing_Walking_031_031b_000269-a_186831018_rgb
|
| 621 |
+
BIWIRGBDID Testing_Walking_031_031b_000287-a_186832933_rgb
|
| 622 |
+
BIWIRGBDID Testing_Walking_031_031b_000341-a_186838350_rgb
|
| 623 |
+
BIWIRGBDID Testing_Walking_031_031b_000353-a_186839350_rgb
|
| 624 |
+
BIWIRGBDID Testing_Walking_031_031b_000361-a_186840101_rgb
|
| 625 |
+
BIWIRGBDID Testing_Walking_031_031b_000368-a_186840685_rgb
|
| 626 |
+
BIWIRGBDID Testing_Walking_031_031b_000371-a_186841017_rgb
|
| 627 |
+
BIWIRGBDID Testing_Walking_031_031b_000380-a_186841933_rgb
|
| 628 |
+
BIWIRGBDID Testing_Walking_031_031b_000400-a_186844101_rgb
|
| 629 |
+
BIWIRGBDID Testing_Walking_031_031b_000402-a_186844349_rgb
|
| 630 |
+
BIWIRGBDID Testing_Walking_031_031b_000403-a_186844433_rgb
|
| 631 |
+
BIWIRGBDID Testing_Walking_032_032b_000022-a_111090932_rgb
|
| 632 |
+
BIWIRGBDID Testing_Walking_032_032b_000034-a_111092016_rgb
|
| 633 |
+
BIWIRGBDID Testing_Walking_032_032b_000038-a_111092516_rgb
|
| 634 |
+
BIWIRGBDID Testing_Walking_032_032b_000044-a_111093100_rgb
|
| 635 |
+
BIWIRGBDID Testing_Walking_032_032b_000053-a_111094100_rgb
|
| 636 |
+
BIWIRGBDID Testing_Walking_032_032b_000127-a_111101515_rgb
|
| 637 |
+
BIWIRGBDID Testing_Walking_032_032b_000128-a_111101599_rgb
|
| 638 |
+
BIWIRGBDID Testing_Walking_032_032b_000154-a_111104015_rgb
|
| 639 |
+
BIWIRGBDID Testing_Walking_032_032b_000160-a_111104684_rgb
|
| 640 |
+
BIWIRGBDID Testing_Walking_032_032b_000184-a_111107100_rgb
|
| 641 |
+
BIWIRGBDID Testing_Walking_032_032b_000187-a_111107515_rgb
|
| 642 |
+
BIWIRGBDID Testing_Walking_032_032b_000254-a_111114099_rgb
|
| 643 |
+
BIWIRGBDID Testing_Walking_032_032b_000302-a_111118431_rgb
|
| 644 |
+
BIWIRGBDID Testing_Walking_032_032b_000308-a_111119015_rgb
|
| 645 |
+
BIWIRGBDID Testing_Walking_032_032b_000313-a_111119599_rgb
|
| 646 |
+
BIWIRGBDID Testing_Walking_032_032b_000324-a_111120763_rgb
|
| 647 |
+
BIWIRGBDID Testing_Walking_032_032b_000332-a_111121683_rgb
|
| 648 |
+
BIWIRGBDID Testing_Walking_032_032b_000335-a_111122015_rgb
|
| 649 |
+
BIWIRGBDID Testing_Walking_032_032b_000351-a_111123683_rgb
|
| 650 |
+
BIWIRGBDID Testing_Walking_032_032b_000363-a_111124763_rgb
|
| 651 |
+
BIWIRGBDID Testing_Walking_032_032b_000385-a_111126847_rgb
|
| 652 |
+
BIWIRGBDID Testing_Walking_032_032b_000404-a_111128763_rgb
|
| 653 |
+
BIWIRGBDID Testing_Walking_036_036b_000040-a_107787409_rgb
|
| 654 |
+
BIWIRGBDID Testing_Walking_036_036b_000049-a_107788409_rgb
|
| 655 |
+
BIWIRGBDID Testing_Walking_036_036b_000121-a_107795493_rgb
|
| 656 |
+
BIWIRGBDID Testing_Walking_036_036b_000150-a_107798077_rgb
|
| 657 |
+
BIWIRGBDID Testing_Walking_036_036b_000152-a_107798241_rgb
|
| 658 |
+
BIWIRGBDID Testing_Walking_036_036b_000176-a_107800661_rgb
|
| 659 |
+
BIWIRGBDID Testing_Walking_036_036b_000181-a_107801241_rgb
|
| 660 |
+
BIWIRGBDID Testing_Walking_036_036b_000196-a_107802825_rgb
|
| 661 |
+
BIWIRGBDID Testing_Walking_036_036b_000210-a_107804241_rgb
|
| 662 |
+
BIWIRGBDID Testing_Walking_036_036b_000213-a_107804577_rgb
|
| 663 |
+
BIWIRGBDID Testing_Walking_036_036b_000215-a_107804824_rgb
|
| 664 |
+
BIWIRGBDID Testing_Walking_036_036b_000220-a_107805325_rgb
|
| 665 |
+
BIWIRGBDID Testing_Walking_036_036b_000226-a_107805908_rgb
|
| 666 |
+
BIWIRGBDID Testing_Walking_036_036b_000253-a_107808324_rgb
|
| 667 |
+
BIWIRGBDID Testing_Walking_036_036b_000277-a_107810657_rgb
|
| 668 |
+
BIWIRGBDID Testing_Walking_036_036b_000292-a_107812240_rgb
|
| 669 |
+
BIWIRGBDID Testing_Walking_036_036b_000303-a_107813324_rgb
|
| 670 |
+
BIWIRGBDID Testing_Walking_036_036b_000344-a_107817741_rgb
|
| 671 |
+
BIWIRGBDID Testing_Walking_036_036b_000373-a_107820408_rgb
|
| 672 |
+
BIWIRGBDID Testing_Walking_036_036b_000387-a_107821656_rgb
|
| 673 |
+
BIWIRGBDID Testing_Walking_037_037b_000029-a_191452363_rgb
|
| 674 |
+
BIWIRGBDID Testing_Walking_037_037b_000054-a_191455364_rgb
|
| 675 |
+
BIWIRGBDID Testing_Walking_037_037b_000058-a_191455779_rgb
|
| 676 |
+
BIWIRGBDID Testing_Walking_037_037b_000075-a_191457527_rgb
|
| 677 |
+
BIWIRGBDID Testing_Walking_037_037b_000127-a_191463363_rgb
|
| 678 |
+
BIWIRGBDID Testing_Walking_037_037b_000146-a_191465279_rgb
|
| 679 |
+
BIWIRGBDID Testing_Walking_037_037b_000166-a_191467527_rgb
|
| 680 |
+
BIWIRGBDID Testing_Walking_037_037b_000169-a_191467863_rgb
|
| 681 |
+
BIWIRGBDID Testing_Walking_037_037b_000179-a_191469027_rgb
|
| 682 |
+
BIWIRGBDID Testing_Walking_037_037b_000188-a_191469947_rgb
|
| 683 |
+
BIWIRGBDID Testing_Walking_037_037b_000215-a_191473110_rgb
|
| 684 |
+
BIWIRGBDID Testing_Walking_037_037b_000227-a_191474362_rgb
|
| 685 |
+
BIWIRGBDID Testing_Walking_037_037b_000251-a_191476694_rgb
|
| 686 |
+
BIWIRGBDID Testing_Walking_037_037b_000254-a_191476942_rgb
|
| 687 |
+
BIWIRGBDID Testing_Walking_037_037b_000259-a_191477362_rgb
|
| 688 |
+
BIWIRGBDID Testing_Walking_037_037b_000270-a_191478694_rgb
|
| 689 |
+
BIWIRGBDID Testing_Walking_037_037b_000277-a_191479363_rgb
|
| 690 |
+
BIWIRGBDID Testing_Walking_037_037b_000291-a_191480695_rgb
|
| 691 |
+
BIWIRGBDID Testing_Walking_037_037b_000299-a_191481694_rgb
|
| 692 |
+
BIWIRGBDID Testing_Walking_037_037b_000308-a_191482778_rgb
|
| 693 |
+
BIWIRGBDID Testing_Walking_037_037b_000322-a_191484194_rgb
|
| 694 |
+
BIWIRGBDID Testing_Walking_037_037b_000323-a_191484362_rgb
|
| 695 |
+
BIWIRGBDID Testing_Walking_037_037b_000345-a_191486610_rgb
|
| 696 |
+
BIWIRGBDID Testing_Walking_039_039b_000002-a_1219323133_rgb
|
| 697 |
+
BIWIRGBDID Testing_Walking_039_039b_000017-a_1219324469_rgb
|
| 698 |
+
BIWIRGBDID Testing_Walking_039_039b_000033-a_1219326048_rgb
|
| 699 |
+
BIWIRGBDID Testing_Walking_039_039b_000075-a_1219330465_rgb
|
| 700 |
+
BIWIRGBDID Testing_Walking_039_039b_000083-a_1219331300_rgb
|
| 701 |
+
BIWIRGBDID Testing_Walking_039_039b_000109-a_1219333632_rgb
|
| 702 |
+
BIWIRGBDID Testing_Walking_039_039b_000123-a_1219334800_rgb
|
| 703 |
+
BIWIRGBDID Testing_Walking_039_039b_000126-a_1219335048_rgb
|
| 704 |
+
BIWIRGBDID Testing_Walking_039_039b_000136-a_1219335884_rgb
|
| 705 |
+
BIWIRGBDID Testing_Walking_039_039b_000150-a_1219337300_rgb
|
| 706 |
+
BIWIRGBDID Testing_Walking_039_039b_000158-a_1219338132_rgb
|
| 707 |
+
BIWIRGBDID Testing_Walking_039_039b_000234-a_1219345216_rgb
|
| 708 |
+
BIWIRGBDID Testing_Walking_039_039b_000253-a_1219346964_rgb
|
| 709 |
+
BIWIRGBDID Testing_Walking_039_039b_000259-a_1219347632_rgb
|
| 710 |
+
BIWIRGBDID Testing_Walking_039_039b_000261-a_1219347880_rgb
|
| 711 |
+
BIWIRGBDID Testing_Walking_039_039b_000274-a_1219349300_rgb
|
| 712 |
+
BIWIRGBDID Testing_Walking_039_039b_000286-a_1219350548_rgb
|
| 713 |
+
BIWIRGBDID Testing_Walking_039_039b_000294-a_1219351380_rgb
|
| 714 |
+
BIWIRGBDID Testing_Walking_039_039b_000295-a_1219351464_rgb
|
| 715 |
+
BIWIRGBDID Testing_Walking_039_039b_000302-a_1219352216_rgb
|
| 716 |
+
BIWIRGBDID Testing_Walking_039_039b_000304-a_1219352380_rgb
|
| 717 |
+
BIWIRGBDID Testing_Walking_039_039b_000305-a_1219352464_rgb
|
| 718 |
+
BIWIRGBDID Testing_Walking_039_039b_000327-a_1219354300_rgb
|
| 719 |
+
BIWIRGBDID Testing_Walking_039_039b_000336-a_1219355048_rgb
|
| 720 |
+
BIWIRGBDID Testing_Walking_039_039b_000375-a_1219358632_rgb
|
| 721 |
+
BIWIRGBDID Testing_Walking_040_040b_000034-a_1231749844_rgb
|
| 722 |
+
BIWIRGBDID Testing_Walking_040_040b_000047-a_1231751260_rgb
|
| 723 |
+
BIWIRGBDID Testing_Walking_040_040b_000063-a_1231752924_rgb
|
| 724 |
+
BIWIRGBDID Testing_Walking_040_040b_000074-a_1231754176_rgb
|
| 725 |
+
BIWIRGBDID Testing_Walking_040_040b_000087-a_1231755508_rgb
|
| 726 |
+
BIWIRGBDID Testing_Walking_040_040b_000121-a_1231759260_rgb
|
| 727 |
+
BIWIRGBDID Testing_Walking_040_040b_000127-a_1231759840_rgb
|
| 728 |
+
BIWIRGBDID Testing_Walking_040_040b_000129-a_1231760092_rgb
|
| 729 |
+
BIWIRGBDID Testing_Walking_040_040b_000138-a_1231761260_rgb
|
| 730 |
+
BIWIRGBDID Testing_Walking_040_040b_000148-a_1231762340_rgb
|
| 731 |
+
BIWIRGBDID Testing_Walking_040_040b_000184-a_1231766260_rgb
|
| 732 |
+
BIWIRGBDID Testing_Walking_040_040b_000250-a_1231773424_rgb
|
| 733 |
+
BIWIRGBDID Testing_Walking_040_040b_000308-a_1231779840_rgb
|
| 734 |
+
BIWIRGBDID Testing_Walking_040_040b_000384-a_1231787923_rgb
|
| 735 |
+
BIWIRGBDID Testing_Walking_044_044b_000036-a_1214190213_rgb
|
| 736 |
+
BIWIRGBDID Testing_Walking_044_044b_000105-a_1214196713_rgb
|
| 737 |
+
BIWIRGBDID Testing_Walking_044_044b_000108-a_1214197045_rgb
|
| 738 |
+
BIWIRGBDID Testing_Walking_044_044b_000134-a_1214199797_rgb
|
| 739 |
+
BIWIRGBDID Testing_Walking_044_044b_000152-a_1214201381_rgb
|
| 740 |
+
BIWIRGBDID Testing_Walking_044_044b_000161-a_1214202213_rgb
|
| 741 |
+
BIWIRGBDID Testing_Walking_044_044b_000262-a_1214211712_rgb
|
| 742 |
+
BIWIRGBDID Testing_Walking_044_044b_000266-a_1214212129_rgb
|
| 743 |
+
BIWIRGBDID Testing_Walking_044_044b_000285-a_1214214128_rgb
|
| 744 |
+
BIWIRGBDID Testing_Walking_044_044b_000309-a_1214216544_rgb
|
| 745 |
+
BIWIRGBDID Testing_Walking_044_044b_000348-a_1214220128_rgb
|
| 746 |
+
BIWIRGBDID Testing_Walking_044_044b_000354-a_1214220712_rgb
|
| 747 |
+
BIWIRGBDID Testing_Walking_046_046b_000002-a_111236092_rgb
|
| 748 |
+
BIWIRGBDID Testing_Walking_046_046b_000023-a_111238008_rgb
|
| 749 |
+
BIWIRGBDID Testing_Walking_046_046b_000032-a_111238840_rgb
|
| 750 |
+
BIWIRGBDID Testing_Walking_046_046b_000057-a_111241172_rgb
|
| 751 |
+
BIWIRGBDID Testing_Walking_046_046b_000080-a_111243672_rgb
|
| 752 |
+
BIWIRGBDID Testing_Walking_046_046b_000103-a_111246256_rgb
|
| 753 |
+
BIWIRGBDID Testing_Walking_046_046b_000120-a_111247756_rgb
|
| 754 |
+
BIWIRGBDID Testing_Walking_046_046b_000128-a_111248508_rgb
|
| 755 |
+
BIWIRGBDID Testing_Walking_046_046b_000168-a_111252508_rgb
|
| 756 |
+
BIWIRGBDID Testing_Walking_046_046b_000171-a_111252840_rgb
|
| 757 |
+
BIWIRGBDID Testing_Walking_046_046b_000242-a_111259588_rgb
|
| 758 |
+
BIWIRGBDID Testing_Walking_046_046b_000281-a_111263672_rgb
|
| 759 |
+
BIWIRGBDID Testing_Walking_046_046b_000323-a_111267587_rgb
|
| 760 |
+
BIWIRGBDID Testing_Walking_046_046b_000335-a_111268755_rgb
|
| 761 |
+
BIWIRGBDID Testing_Walking_046_046b_000342-a_111269507_rgb
|
| 762 |
+
BIWIRGBDID Testing_Walking_046_046b_000344-a_111269756_rgb
|
| 763 |
+
BIWIRGBDID Testing_Walking_047_047b_000014-a_1214112719_rgb
|
| 764 |
+
BIWIRGBDID Testing_Walking_047_047b_000024-a_1214113551_rgb
|
| 765 |
+
BIWIRGBDID Testing_Walking_047_047b_000035-a_1214114635_rgb
|
| 766 |
+
BIWIRGBDID Testing_Walking_047_047b_000070-a_1214118467_rgb
|
| 767 |
+
BIWIRGBDID Testing_Walking_047_047b_000083-a_1214119719_rgb
|
| 768 |
+
BIWIRGBDID Testing_Walking_047_047b_000094-a_1214120719_rgb
|
| 769 |
+
BIWIRGBDID Testing_Walking_047_047b_000105-a_1214122135_rgb
|
| 770 |
+
BIWIRGBDID Testing_Walking_047_047b_000128-a_1214124635_rgb
|
| 771 |
+
BIWIRGBDID Testing_Walking_047_047b_000144-a_1214126551_rgb
|
| 772 |
+
BIWIRGBDID Testing_Walking_047_047b_000146-a_1214126718_rgb
|
| 773 |
+
BIWIRGBDID Testing_Walking_047_047b_000166-a_1214128634_rgb
|
| 774 |
+
BIWIRGBDID Testing_Walking_047_047b_000186-a_1214130303_rgb
|
| 775 |
+
BIWIRGBDID Testing_Walking_047_047b_000197-a_1214131298_rgb
|
| 776 |
+
BIWIRGBDID Testing_Walking_047_047b_000207-a_1214132218_rgb
|
| 777 |
+
BIWIRGBDID Testing_Walking_047_047b_000209-a_1214132382_rgb
|
| 778 |
+
BIWIRGBDID Testing_Walking_047_047b_000213-a_1214132718_rgb
|
| 779 |
+
BIWIRGBDID Testing_Walking_047_047b_000228-a_1214134299_rgb
|
| 780 |
+
BIWIRGBDID Testing_Walking_047_047b_000261-a_1214138134_rgb
|
| 781 |
+
BIWIRGBDID Testing_Walking_047_047b_000303-a_1214142218_rgb
|
| 782 |
+
BIWIRGBDID Testing_Walking_047_047b_000308-a_1214142882_rgb
|
| 783 |
+
BIWIRGBDID Testing_Walking_047_047b_000314-a_1214143466_rgb
|
| 784 |
+
BIWIRGBDID Testing_Walking_047_047b_000336-a_1214145718_rgb
|
| 785 |
+
BIWIRGBDID Testing_Walking_049_049b_000002-a_201380445_rgb
|
| 786 |
+
BIWIRGBDID Testing_Walking_049_049b_000132-a_201394025_rgb
|
| 787 |
+
BIWIRGBDID Testing_Walking_049_049b_000138-a_201394693_rgb
|
| 788 |
+
BIWIRGBDID Testing_Walking_049_049b_000145-a_201395360_rgb
|
| 789 |
+
BIWIRGBDID Testing_Walking_049_049b_000166-a_201397609_rgb
|
| 790 |
+
BIWIRGBDID Testing_Walking_049_049b_000257-a_201407193_rgb
|
| 791 |
+
BIWIRGBDID Testing_Walking_049_049b_000272-a_201408776_rgb
|
| 792 |
+
BIWIRGBDID Testing_Walking_049_049b_000287-a_201410440_rgb
|
| 793 |
+
BIWIRGBDID Testing_Walking_049_049b_000295-a_201411276_rgb
|
| 794 |
+
BIWIRGBDID Testing_Walking_049_049b_000312-a_201413192_rgb
|
| 795 |
+
BIWIRGBDID Testing_Walking_049_049b_000316-a_201413608_rgb
|
| 796 |
+
BIWIRGBDID Testing_Walking_049_049b_000335-a_201415608_rgb
|
| 797 |
+
BIWIRGBDID Testing_Walking_049_049b_000366-a_201418940_rgb
|
| 798 |
+
BIWIRGBDID Testing_Walking_049_049b_000395-a_201422108_rgb
|
| 799 |
+
BIWIRGBDID Testing_Walking_049_049b_000409-a_201423524_rgb
|
| 800 |
+
BIWIRGBDID Testing_Walking_049_049b_000429-a_201425692_rgb
|
external/WildCamera/splits/cad120_test.txt
ADDED
|
@@ -0,0 +1,800 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 1.jpg
|
| 2 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 11.jpg
|
| 3 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 61.jpg
|
| 4 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 71.jpg
|
| 5 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 121.jpg
|
| 6 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 131.jpg
|
| 7 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 141.jpg
|
| 8 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 151.jpg
|
| 9 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 161.jpg
|
| 10 |
+
CAD120_Subject1_rgbd_images_arranging_objects_sequence 171.jpg
|
| 11 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 1.jpg
|
| 12 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 41.jpg
|
| 13 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 61.jpg
|
| 14 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 71.jpg
|
| 15 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 81.jpg
|
| 16 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 101.jpg
|
| 17 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 111.jpg
|
| 18 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 121.jpg
|
| 19 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 131.jpg
|
| 20 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 141.jpg
|
| 21 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 151.jpg
|
| 22 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 161.jpg
|
| 23 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 171.jpg
|
| 24 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 181.jpg
|
| 25 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 201.jpg
|
| 26 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 221.jpg
|
| 27 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 241.jpg
|
| 28 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 291.jpg
|
| 29 |
+
CAD120_Subject1_rgbd_images_cleaning_objects_sequence 321.jpg
|
| 30 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 1.jpg
|
| 31 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 71.jpg
|
| 32 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 81.jpg
|
| 33 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 101.jpg
|
| 34 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 141.jpg
|
| 35 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 161.jpg
|
| 36 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 171.jpg
|
| 37 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 201.jpg
|
| 38 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 211.jpg
|
| 39 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 221.jpg
|
| 40 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 231.jpg
|
| 41 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 241.jpg
|
| 42 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 281.jpg
|
| 43 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 291.jpg
|
| 44 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 311.jpg
|
| 45 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 331.jpg
|
| 46 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 341.jpg
|
| 47 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 361.jpg
|
| 48 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 381.jpg
|
| 49 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 391.jpg
|
| 50 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 401.jpg
|
| 51 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 411.jpg
|
| 52 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 421.jpg
|
| 53 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 431.jpg
|
| 54 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 441.jpg
|
| 55 |
+
CAD120_Subject1_rgbd_images_having_meal_sequence 471.jpg
|
| 56 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 1.jpg
|
| 57 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 11.jpg
|
| 58 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 41.jpg
|
| 59 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 51.jpg
|
| 60 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 81.jpg
|
| 61 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 91.jpg
|
| 62 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 101.jpg
|
| 63 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 111.jpg
|
| 64 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 131.jpg
|
| 65 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 151.jpg
|
| 66 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 161.jpg
|
| 67 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 171.jpg
|
| 68 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 211.jpg
|
| 69 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 221.jpg
|
| 70 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 231.jpg
|
| 71 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 261.jpg
|
| 72 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 281.jpg
|
| 73 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 301.jpg
|
| 74 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 321.jpg
|
| 75 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 351.jpg
|
| 76 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 371.jpg
|
| 77 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 381.jpg
|
| 78 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 401.jpg
|
| 79 |
+
CAD120_Subject1_rgbd_images_making_cereal_sequence 421.jpg
|
| 80 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 1.jpg
|
| 81 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 11.jpg
|
| 82 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 21.jpg
|
| 83 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 31.jpg
|
| 84 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 41.jpg
|
| 85 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 91.jpg
|
| 86 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 101.jpg
|
| 87 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 111.jpg
|
| 88 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 121.jpg
|
| 89 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 131.jpg
|
| 90 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 141.jpg
|
| 91 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 191.jpg
|
| 92 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 201.jpg
|
| 93 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 221.jpg
|
| 94 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 231.jpg
|
| 95 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 241.jpg
|
| 96 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 261.jpg
|
| 97 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 281.jpg
|
| 98 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 291.jpg
|
| 99 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 301.jpg
|
| 100 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 321.jpg
|
| 101 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 341.jpg
|
| 102 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 351.jpg
|
| 103 |
+
CAD120_Subject1_rgbd_images_microwaving_food_sequence 381.jpg
|
| 104 |
+
CAD120_Subject1_rgbd_images_picking_objects_sequence 1.jpg
|
| 105 |
+
CAD120_Subject1_rgbd_images_picking_objects_sequence 31.jpg
|
| 106 |
+
CAD120_Subject1_rgbd_images_picking_objects_sequence 41.jpg
|
| 107 |
+
CAD120_Subject1_rgbd_images_picking_objects_sequence 71.jpg
|
| 108 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 1.jpg
|
| 109 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 11.jpg
|
| 110 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 21.jpg
|
| 111 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 31.jpg
|
| 112 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 51.jpg
|
| 113 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 61.jpg
|
| 114 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 71.jpg
|
| 115 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 91.jpg
|
| 116 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 101.jpg
|
| 117 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 121.jpg
|
| 118 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 131.jpg
|
| 119 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 141.jpg
|
| 120 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 161.jpg
|
| 121 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 171.jpg
|
| 122 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 211.jpg
|
| 123 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 221.jpg
|
| 124 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 231.jpg
|
| 125 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 241.jpg
|
| 126 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 251.jpg
|
| 127 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 271.jpg
|
| 128 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 281.jpg
|
| 129 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 301.jpg
|
| 130 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 321.jpg
|
| 131 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 341.jpg
|
| 132 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 351.jpg
|
| 133 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 361.jpg
|
| 134 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 371.jpg
|
| 135 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 381.jpg
|
| 136 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 401.jpg
|
| 137 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 411.jpg
|
| 138 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 421.jpg
|
| 139 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 441.jpg
|
| 140 |
+
CAD120_Subject1_rgbd_images_stacking_objects_sequence 461.jpg
|
| 141 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 11.jpg
|
| 142 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 21.jpg
|
| 143 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 31.jpg
|
| 144 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 41.jpg
|
| 145 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 71.jpg
|
| 146 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 81.jpg
|
| 147 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 91.jpg
|
| 148 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 111.jpg
|
| 149 |
+
CAD120_Subject1_rgbd_images_taking_food_sequence 131.jpg
|
| 150 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 21.jpg
|
| 151 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 31.jpg
|
| 152 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 41.jpg
|
| 153 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 51.jpg
|
| 154 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 71.jpg
|
| 155 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 81.jpg
|
| 156 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 91.jpg
|
| 157 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 101.jpg
|
| 158 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 111.jpg
|
| 159 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 121.jpg
|
| 160 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 131.jpg
|
| 161 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 141.jpg
|
| 162 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 151.jpg
|
| 163 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 171.jpg
|
| 164 |
+
CAD120_Subject1_rgbd_images_taking_medicine_sequence 201.jpg
|
| 165 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 1.jpg
|
| 166 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 11.jpg
|
| 167 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 51.jpg
|
| 168 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 61.jpg
|
| 169 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 71.jpg
|
| 170 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 81.jpg
|
| 171 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 111.jpg
|
| 172 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 141.jpg
|
| 173 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 151.jpg
|
| 174 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 161.jpg
|
| 175 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 181.jpg
|
| 176 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 201.jpg
|
| 177 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 221.jpg
|
| 178 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 231.jpg
|
| 179 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 241.jpg
|
| 180 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 261.jpg
|
| 181 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 291.jpg
|
| 182 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 311.jpg
|
| 183 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 351.jpg
|
| 184 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 361.jpg
|
| 185 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 371.jpg
|
| 186 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 391.jpg
|
| 187 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 401.jpg
|
| 188 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 461.jpg
|
| 189 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 471.jpg
|
| 190 |
+
CAD120_Subject1_rgbd_images_unstacking_objects_sequence 501.jpg
|
| 191 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 1.jpg
|
| 192 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 11.jpg
|
| 193 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 21.jpg
|
| 194 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 31.jpg
|
| 195 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 41.jpg
|
| 196 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 51.jpg
|
| 197 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 61.jpg
|
| 198 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 71.jpg
|
| 199 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 81.jpg
|
| 200 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 101.jpg
|
| 201 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 111.jpg
|
| 202 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 131.jpg
|
| 203 |
+
CAD120_Subject3_rgbd_images_arranging_objects_sequence 141.jpg
|
| 204 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 1.jpg
|
| 205 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 21.jpg
|
| 206 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 31.jpg
|
| 207 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 41.jpg
|
| 208 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 61.jpg
|
| 209 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 71.jpg
|
| 210 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 101.jpg
|
| 211 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 141.jpg
|
| 212 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 161.jpg
|
| 213 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 181.jpg
|
| 214 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 191.jpg
|
| 215 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 211.jpg
|
| 216 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 221.jpg
|
| 217 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 241.jpg
|
| 218 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 251.jpg
|
| 219 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 261.jpg
|
| 220 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 271.jpg
|
| 221 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 281.jpg
|
| 222 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 291.jpg
|
| 223 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 311.jpg
|
| 224 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 331.jpg
|
| 225 |
+
CAD120_Subject3_rgbd_images_cleaning_objects_sequence 341.jpg
|
| 226 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 1.jpg
|
| 227 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 11.jpg
|
| 228 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 21.jpg
|
| 229 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 51.jpg
|
| 230 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 61.jpg
|
| 231 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 91.jpg
|
| 232 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 101.jpg
|
| 233 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 111.jpg
|
| 234 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 121.jpg
|
| 235 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 131.jpg
|
| 236 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 151.jpg
|
| 237 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 201.jpg
|
| 238 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 221.jpg
|
| 239 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 241.jpg
|
| 240 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 261.jpg
|
| 241 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 271.jpg
|
| 242 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 301.jpg
|
| 243 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 321.jpg
|
| 244 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 341.jpg
|
| 245 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 371.jpg
|
| 246 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 381.jpg
|
| 247 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 411.jpg
|
| 248 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 421.jpg
|
| 249 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 441.jpg
|
| 250 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 461.jpg
|
| 251 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 471.jpg
|
| 252 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 481.jpg
|
| 253 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 491.jpg
|
| 254 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 501.jpg
|
| 255 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 521.jpg
|
| 256 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 531.jpg
|
| 257 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 551.jpg
|
| 258 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 611.jpg
|
| 259 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 621.jpg
|
| 260 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 651.jpg
|
| 261 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 661.jpg
|
| 262 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 671.jpg
|
| 263 |
+
CAD120_Subject3_rgbd_images_having_meal_sequence 691.jpg
|
| 264 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 41.jpg
|
| 265 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 51.jpg
|
| 266 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 61.jpg
|
| 267 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 71.jpg
|
| 268 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 111.jpg
|
| 269 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 141.jpg
|
| 270 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 151.jpg
|
| 271 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 161.jpg
|
| 272 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 171.jpg
|
| 273 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 211.jpg
|
| 274 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 231.jpg
|
| 275 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 241.jpg
|
| 276 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 291.jpg
|
| 277 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 301.jpg
|
| 278 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 321.jpg
|
| 279 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 341.jpg
|
| 280 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 351.jpg
|
| 281 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 361.jpg
|
| 282 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 371.jpg
|
| 283 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 411.jpg
|
| 284 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 421.jpg
|
| 285 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 431.jpg
|
| 286 |
+
CAD120_Subject3_rgbd_images_making_cereal_sequence 441.jpg
|
| 287 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 1.jpg
|
| 288 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 11.jpg
|
| 289 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 41.jpg
|
| 290 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 51.jpg
|
| 291 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 91.jpg
|
| 292 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 101.jpg
|
| 293 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 121.jpg
|
| 294 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 141.jpg
|
| 295 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 151.jpg
|
| 296 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 161.jpg
|
| 297 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 191.jpg
|
| 298 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 211.jpg
|
| 299 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 221.jpg
|
| 300 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 231.jpg
|
| 301 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 241.jpg
|
| 302 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 251.jpg
|
| 303 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 301.jpg
|
| 304 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 311.jpg
|
| 305 |
+
CAD120_Subject3_rgbd_images_microwaving_food_sequence 331.jpg
|
| 306 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 1.jpg
|
| 307 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 11.jpg
|
| 308 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 71.jpg
|
| 309 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 81.jpg
|
| 310 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 101.jpg
|
| 311 |
+
CAD120_Subject3_rgbd_images_picking_objects_sequence 111.jpg
|
| 312 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 1.jpg
|
| 313 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 11.jpg
|
| 314 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 41.jpg
|
| 315 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 51.jpg
|
| 316 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 61.jpg
|
| 317 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 71.jpg
|
| 318 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 81.jpg
|
| 319 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 91.jpg
|
| 320 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 101.jpg
|
| 321 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 111.jpg
|
| 322 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 161.jpg
|
| 323 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 171.jpg
|
| 324 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 181.jpg
|
| 325 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 191.jpg
|
| 326 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 201.jpg
|
| 327 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 221.jpg
|
| 328 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 231.jpg
|
| 329 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 241.jpg
|
| 330 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 251.jpg
|
| 331 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 291.jpg
|
| 332 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 321.jpg
|
| 333 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 331.jpg
|
| 334 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 371.jpg
|
| 335 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 381.jpg
|
| 336 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 391.jpg
|
| 337 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 401.jpg
|
| 338 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 411.jpg
|
| 339 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 421.jpg
|
| 340 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 431.jpg
|
| 341 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 471.jpg
|
| 342 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 481.jpg
|
| 343 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 491.jpg
|
| 344 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 501.jpg
|
| 345 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 521.jpg
|
| 346 |
+
CAD120_Subject3_rgbd_images_stacking_objects_sequence 551.jpg
|
| 347 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 21.jpg
|
| 348 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 31.jpg
|
| 349 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 51.jpg
|
| 350 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 81.jpg
|
| 351 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 101.jpg
|
| 352 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 111.jpg
|
| 353 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 141.jpg
|
| 354 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 161.jpg
|
| 355 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 181.jpg
|
| 356 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 201.jpg
|
| 357 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 211.jpg
|
| 358 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 231.jpg
|
| 359 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 241.jpg
|
| 360 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 251.jpg
|
| 361 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 271.jpg
|
| 362 |
+
CAD120_Subject3_rgbd_images_taking_food_sequence 291.jpg
|
| 363 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 1.jpg
|
| 364 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 11.jpg
|
| 365 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 21.jpg
|
| 366 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 31.jpg
|
| 367 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 61.jpg
|
| 368 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 71.jpg
|
| 369 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 101.jpg
|
| 370 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 111.jpg
|
| 371 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 131.jpg
|
| 372 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 151.jpg
|
| 373 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 211.jpg
|
| 374 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 231.jpg
|
| 375 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 261.jpg
|
| 376 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 301.jpg
|
| 377 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 311.jpg
|
| 378 |
+
CAD120_Subject3_rgbd_images_taking_medicine_sequence 321.jpg
|
| 379 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 11.jpg
|
| 380 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 21.jpg
|
| 381 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 31.jpg
|
| 382 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 41.jpg
|
| 383 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 61.jpg
|
| 384 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 71.jpg
|
| 385 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 81.jpg
|
| 386 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 101.jpg
|
| 387 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 111.jpg
|
| 388 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 131.jpg
|
| 389 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 141.jpg
|
| 390 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 151.jpg
|
| 391 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 171.jpg
|
| 392 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 191.jpg
|
| 393 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 221.jpg
|
| 394 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 231.jpg
|
| 395 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 251.jpg
|
| 396 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 261.jpg
|
| 397 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 281.jpg
|
| 398 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 301.jpg
|
| 399 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 321.jpg
|
| 400 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 341.jpg
|
| 401 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 351.jpg
|
| 402 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 451.jpg
|
| 403 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 471.jpg
|
| 404 |
+
CAD120_Subject3_rgbd_images_unstacking_objects_sequence 501.jpg
|
| 405 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 1.jpg
|
| 406 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 11.jpg
|
| 407 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 21.jpg
|
| 408 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 31.jpg
|
| 409 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 71.jpg
|
| 410 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 91.jpg
|
| 411 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 101.jpg
|
| 412 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 111.jpg
|
| 413 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 121.jpg
|
| 414 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 131.jpg
|
| 415 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 141.jpg
|
| 416 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 161.jpg
|
| 417 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 171.jpg
|
| 418 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 181.jpg
|
| 419 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 201.jpg
|
| 420 |
+
CAD120_Subject4_rgbd_images_arranging_objects_sequence 211.jpg
|
| 421 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 11.jpg
|
| 422 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 21.jpg
|
| 423 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 31.jpg
|
| 424 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 51.jpg
|
| 425 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 71.jpg
|
| 426 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 81.jpg
|
| 427 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 141.jpg
|
| 428 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 171.jpg
|
| 429 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 191.jpg
|
| 430 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 201.jpg
|
| 431 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 211.jpg
|
| 432 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 221.jpg
|
| 433 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 241.jpg
|
| 434 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 251.jpg
|
| 435 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 261.jpg
|
| 436 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 271.jpg
|
| 437 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 291.jpg
|
| 438 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 301.jpg
|
| 439 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 331.jpg
|
| 440 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 341.jpg
|
| 441 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 351.jpg
|
| 442 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 371.jpg
|
| 443 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 381.jpg
|
| 444 |
+
CAD120_Subject4_rgbd_images_cleaning_objects_sequence 401.jpg
|
| 445 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 1.jpg
|
| 446 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 21.jpg
|
| 447 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 41.jpg
|
| 448 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 71.jpg
|
| 449 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 101.jpg
|
| 450 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 121.jpg
|
| 451 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 141.jpg
|
| 452 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 151.jpg
|
| 453 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 161.jpg
|
| 454 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 171.jpg
|
| 455 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 181.jpg
|
| 456 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 201.jpg
|
| 457 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 221.jpg
|
| 458 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 231.jpg
|
| 459 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 241.jpg
|
| 460 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 251.jpg
|
| 461 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 271.jpg
|
| 462 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 311.jpg
|
| 463 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 331.jpg
|
| 464 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 341.jpg
|
| 465 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 351.jpg
|
| 466 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 361.jpg
|
| 467 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 371.jpg
|
| 468 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 401.jpg
|
| 469 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 411.jpg
|
| 470 |
+
CAD120_Subject4_rgbd_images_having_meal_sequence 461.jpg
|
| 471 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 1.jpg
|
| 472 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 31.jpg
|
| 473 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 61.jpg
|
| 474 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 71.jpg
|
| 475 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 81.jpg
|
| 476 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 91.jpg
|
| 477 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 111.jpg
|
| 478 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 121.jpg
|
| 479 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 151.jpg
|
| 480 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 191.jpg
|
| 481 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 211.jpg
|
| 482 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 221.jpg
|
| 483 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 231.jpg
|
| 484 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 241.jpg
|
| 485 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 251.jpg
|
| 486 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 271.jpg
|
| 487 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 281.jpg
|
| 488 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 291.jpg
|
| 489 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 301.jpg
|
| 490 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 321.jpg
|
| 491 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 341.jpg
|
| 492 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 351.jpg
|
| 493 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 371.jpg
|
| 494 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 381.jpg
|
| 495 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 391.jpg
|
| 496 |
+
CAD120_Subject4_rgbd_images_making_cereal_sequence 401.jpg
|
| 497 |
+
CAD120_Subject4_rgbd_images_microwaving_food_sequence 11.jpg
|
| 498 |
+
CAD120_Subject4_rgbd_images_microwaving_food_sequence 21.jpg
|
| 499 |
+
CAD120_Subject4_rgbd_images_microwaving_food_sequence 61.jpg
|
| 500 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 11.jpg
|
| 501 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 21.jpg
|
| 502 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 31.jpg
|
| 503 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 61.jpg
|
| 504 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 71.jpg
|
| 505 |
+
CAD120_Subject4_rgbd_images_picking_objects_sequence 91.jpg
|
| 506 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 1.jpg
|
| 507 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 11.jpg
|
| 508 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 21.jpg
|
| 509 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 31.jpg
|
| 510 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 61.jpg
|
| 511 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 71.jpg
|
| 512 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 111.jpg
|
| 513 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 151.jpg
|
| 514 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 161.jpg
|
| 515 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 171.jpg
|
| 516 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 181.jpg
|
| 517 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 191.jpg
|
| 518 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 211.jpg
|
| 519 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 221.jpg
|
| 520 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 231.jpg
|
| 521 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 241.jpg
|
| 522 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 251.jpg
|
| 523 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 271.jpg
|
| 524 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 291.jpg
|
| 525 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 311.jpg
|
| 526 |
+
CAD120_Subject4_rgbd_images_stacking_objects_sequence 321.jpg
|
| 527 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 1.jpg
|
| 528 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 11.jpg
|
| 529 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 21.jpg
|
| 530 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 41.jpg
|
| 531 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 71.jpg
|
| 532 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 91.jpg
|
| 533 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 101.jpg
|
| 534 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 141.jpg
|
| 535 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 151.jpg
|
| 536 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 161.jpg
|
| 537 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 171.jpg
|
| 538 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 181.jpg
|
| 539 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 201.jpg
|
| 540 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 241.jpg
|
| 541 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 261.jpg
|
| 542 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 281.jpg
|
| 543 |
+
CAD120_Subject4_rgbd_images_taking_food_sequence 311.jpg
|
| 544 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 31.jpg
|
| 545 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 51.jpg
|
| 546 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 61.jpg
|
| 547 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 71.jpg
|
| 548 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 81.jpg
|
| 549 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 91.jpg
|
| 550 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 121.jpg
|
| 551 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 141.jpg
|
| 552 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 151.jpg
|
| 553 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 171.jpg
|
| 554 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 181.jpg
|
| 555 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 191.jpg
|
| 556 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 221.jpg
|
| 557 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 231.jpg
|
| 558 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 241.jpg
|
| 559 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 261.jpg
|
| 560 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 271.jpg
|
| 561 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 281.jpg
|
| 562 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 291.jpg
|
| 563 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 301.jpg
|
| 564 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 321.jpg
|
| 565 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 331.jpg
|
| 566 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 341.jpg
|
| 567 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 371.jpg
|
| 568 |
+
CAD120_Subject4_rgbd_images_taking_medicine_sequence 381.jpg
|
| 569 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 1.jpg
|
| 570 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 31.jpg
|
| 571 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 51.jpg
|
| 572 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 61.jpg
|
| 573 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 71.jpg
|
| 574 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 81.jpg
|
| 575 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 91.jpg
|
| 576 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 111.jpg
|
| 577 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 131.jpg
|
| 578 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 141.jpg
|
| 579 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 151.jpg
|
| 580 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 161.jpg
|
| 581 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 171.jpg
|
| 582 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 191.jpg
|
| 583 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 201.jpg
|
| 584 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 221.jpg
|
| 585 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 241.jpg
|
| 586 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 251.jpg
|
| 587 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 261.jpg
|
| 588 |
+
CAD120_Subject4_rgbd_images_unstacking_objects_sequence 271.jpg
|
| 589 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 11.jpg
|
| 590 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 21.jpg
|
| 591 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 31.jpg
|
| 592 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 41.jpg
|
| 593 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 71.jpg
|
| 594 |
+
CAD120_Subject5_rgbd_images_arranging_objects_sequence 81.jpg
|
| 595 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 21.jpg
|
| 596 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 31.jpg
|
| 597 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 51.jpg
|
| 598 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 71.jpg
|
| 599 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 81.jpg
|
| 600 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 101.jpg
|
| 601 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 111.jpg
|
| 602 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 121.jpg
|
| 603 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 131.jpg
|
| 604 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 141.jpg
|
| 605 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 151.jpg
|
| 606 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 171.jpg
|
| 607 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 201.jpg
|
| 608 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 211.jpg
|
| 609 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 221.jpg
|
| 610 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 231.jpg
|
| 611 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 251.jpg
|
| 612 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 271.jpg
|
| 613 |
+
CAD120_Subject5_rgbd_images_cleaning_objects_sequence 281.jpg
|
| 614 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 1.jpg
|
| 615 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 11.jpg
|
| 616 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 21.jpg
|
| 617 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 31.jpg
|
| 618 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 41.jpg
|
| 619 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 61.jpg
|
| 620 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 101.jpg
|
| 621 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 111.jpg
|
| 622 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 151.jpg
|
| 623 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 201.jpg
|
| 624 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 221.jpg
|
| 625 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 251.jpg
|
| 626 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 271.jpg
|
| 627 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 281.jpg
|
| 628 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 351.jpg
|
| 629 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 361.jpg
|
| 630 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 371.jpg
|
| 631 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 381.jpg
|
| 632 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 421.jpg
|
| 633 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 431.jpg
|
| 634 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 441.jpg
|
| 635 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 451.jpg
|
| 636 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 461.jpg
|
| 637 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 471.jpg
|
| 638 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 481.jpg
|
| 639 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 531.jpg
|
| 640 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 541.jpg
|
| 641 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 551.jpg
|
| 642 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 561.jpg
|
| 643 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 581.jpg
|
| 644 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 621.jpg
|
| 645 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 641.jpg
|
| 646 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 651.jpg
|
| 647 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 671.jpg
|
| 648 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 681.jpg
|
| 649 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 691.jpg
|
| 650 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 701.jpg
|
| 651 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 711.jpg
|
| 652 |
+
CAD120_Subject5_rgbd_images_having_meal_sequence 721.jpg
|
| 653 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 11.jpg
|
| 654 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 31.jpg
|
| 655 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 41.jpg
|
| 656 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 71.jpg
|
| 657 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 91.jpg
|
| 658 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 101.jpg
|
| 659 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 121.jpg
|
| 660 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 161.jpg
|
| 661 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 171.jpg
|
| 662 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 181.jpg
|
| 663 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 191.jpg
|
| 664 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 221.jpg
|
| 665 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 241.jpg
|
| 666 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 271.jpg
|
| 667 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 281.jpg
|
| 668 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 301.jpg
|
| 669 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 331.jpg
|
| 670 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 341.jpg
|
| 671 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 351.jpg
|
| 672 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 361.jpg
|
| 673 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 371.jpg
|
| 674 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 381.jpg
|
| 675 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 391.jpg
|
| 676 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 411.jpg
|
| 677 |
+
CAD120_Subject5_rgbd_images_making_cereal_sequence 441.jpg
|
| 678 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 1.jpg
|
| 679 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 11.jpg
|
| 680 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 21.jpg
|
| 681 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 31.jpg
|
| 682 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 41.jpg
|
| 683 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 61.jpg
|
| 684 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 81.jpg
|
| 685 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 91.jpg
|
| 686 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 101.jpg
|
| 687 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 111.jpg
|
| 688 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 121.jpg
|
| 689 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 131.jpg
|
| 690 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 161.jpg
|
| 691 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 171.jpg
|
| 692 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 181.jpg
|
| 693 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 191.jpg
|
| 694 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 231.jpg
|
| 695 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 241.jpg
|
| 696 |
+
CAD120_Subject5_rgbd_images_microwaving_food_sequence 271.jpg
|
| 697 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 1.jpg
|
| 698 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 11.jpg
|
| 699 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 21.jpg
|
| 700 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 51.jpg
|
| 701 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 81.jpg
|
| 702 |
+
CAD120_Subject5_rgbd_images_picking_objects_sequence 121.jpg
|
| 703 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 1.jpg
|
| 704 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 11.jpg
|
| 705 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 41.jpg
|
| 706 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 51.jpg
|
| 707 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 61.jpg
|
| 708 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 81.jpg
|
| 709 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 91.jpg
|
| 710 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 101.jpg
|
| 711 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 111.jpg
|
| 712 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 141.jpg
|
| 713 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 161.jpg
|
| 714 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 191.jpg
|
| 715 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 281.jpg
|
| 716 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 291.jpg
|
| 717 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 301.jpg
|
| 718 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 311.jpg
|
| 719 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 331.jpg
|
| 720 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 351.jpg
|
| 721 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 361.jpg
|
| 722 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 371.jpg
|
| 723 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 381.jpg
|
| 724 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 391.jpg
|
| 725 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 401.jpg
|
| 726 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 421.jpg
|
| 727 |
+
CAD120_Subject5_rgbd_images_stacking_objects_sequence 431.jpg
|
| 728 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 11.jpg
|
| 729 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 51.jpg
|
| 730 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 71.jpg
|
| 731 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 81.jpg
|
| 732 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 101.jpg
|
| 733 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 111.jpg
|
| 734 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 121.jpg
|
| 735 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 131.jpg
|
| 736 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 141.jpg
|
| 737 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 171.jpg
|
| 738 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 191.jpg
|
| 739 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 211.jpg
|
| 740 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 221.jpg
|
| 741 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 231.jpg
|
| 742 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 241.jpg
|
| 743 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 261.jpg
|
| 744 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 311.jpg
|
| 745 |
+
CAD120_Subject5_rgbd_images_taking_food_sequence 321.jpg
|
| 746 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 1.jpg
|
| 747 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 41.jpg
|
| 748 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 51.jpg
|
| 749 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 61.jpg
|
| 750 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 81.jpg
|
| 751 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 91.jpg
|
| 752 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 101.jpg
|
| 753 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 111.jpg
|
| 754 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 121.jpg
|
| 755 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 131.jpg
|
| 756 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 151.jpg
|
| 757 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 161.jpg
|
| 758 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 171.jpg
|
| 759 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 181.jpg
|
| 760 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 211.jpg
|
| 761 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 231.jpg
|
| 762 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 241.jpg
|
| 763 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 251.jpg
|
| 764 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 271.jpg
|
| 765 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 281.jpg
|
| 766 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 311.jpg
|
| 767 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 331.jpg
|
| 768 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 341.jpg
|
| 769 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 351.jpg
|
| 770 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 361.jpg
|
| 771 |
+
CAD120_Subject5_rgbd_images_taking_medicine_sequence 391.jpg
|
| 772 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 11.jpg
|
| 773 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 51.jpg
|
| 774 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 61.jpg
|
| 775 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 71.jpg
|
| 776 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 151.jpg
|
| 777 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 161.jpg
|
| 778 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 171.jpg
|
| 779 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 181.jpg
|
| 780 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 191.jpg
|
| 781 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 201.jpg
|
| 782 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 211.jpg
|
| 783 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 241.jpg
|
| 784 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 261.jpg
|
| 785 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 271.jpg
|
| 786 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 281.jpg
|
| 787 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 321.jpg
|
| 788 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 331.jpg
|
| 789 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 371.jpg
|
| 790 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 381.jpg
|
| 791 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 401.jpg
|
| 792 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 411.jpg
|
| 793 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 421.jpg
|
| 794 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 431.jpg
|
| 795 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 441.jpg
|
| 796 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 461.jpg
|
| 797 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 481.jpg
|
| 798 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 491.jpg
|
| 799 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 511.jpg
|
| 800 |
+
CAD120_Subject5_rgbd_images_unstacking_objects_sequence 521.jpg
|
external/WildCamera/splits/cityscapes_test.txt
ADDED
|
@@ -0,0 +1,800 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
berlin berlin_000000_000019
|
| 2 |
+
berlin berlin_000001_000019
|
| 3 |
+
berlin berlin_000002_000019
|
| 4 |
+
berlin berlin_000004_000019
|
| 5 |
+
berlin berlin_000005_000019
|
| 6 |
+
berlin berlin_000009_000019
|
| 7 |
+
berlin berlin_000013_000019
|
| 8 |
+
berlin berlin_000014_000019
|
| 9 |
+
berlin berlin_000016_000019
|
| 10 |
+
berlin berlin_000018_000019
|
| 11 |
+
berlin berlin_000019_000019
|
| 12 |
+
berlin berlin_000020_000019
|
| 13 |
+
berlin berlin_000021_000019
|
| 14 |
+
berlin berlin_000022_000019
|
| 15 |
+
berlin berlin_000023_000019
|
| 16 |
+
berlin berlin_000026_000019
|
| 17 |
+
berlin berlin_000027_000019
|
| 18 |
+
berlin berlin_000029_000019
|
| 19 |
+
berlin berlin_000033_000019
|
| 20 |
+
berlin berlin_000036_000019
|
| 21 |
+
berlin berlin_000037_000019
|
| 22 |
+
berlin berlin_000038_000019
|
| 23 |
+
berlin berlin_000039_000019
|
| 24 |
+
berlin berlin_000040_000019
|
| 25 |
+
berlin berlin_000041_000019
|
| 26 |
+
berlin berlin_000042_000019
|
| 27 |
+
berlin berlin_000044_000019
|
| 28 |
+
berlin berlin_000046_000019
|
| 29 |
+
berlin berlin_000051_000019
|
| 30 |
+
berlin berlin_000052_000019
|
| 31 |
+
berlin berlin_000053_000019
|
| 32 |
+
berlin berlin_000056_000019
|
| 33 |
+
berlin berlin_000058_000019
|
| 34 |
+
berlin berlin_000063_000019
|
| 35 |
+
berlin berlin_000066_000019
|
| 36 |
+
berlin berlin_000068_000019
|
| 37 |
+
berlin berlin_000069_000019
|
| 38 |
+
berlin berlin_000072_000019
|
| 39 |
+
berlin berlin_000073_000019
|
| 40 |
+
berlin berlin_000074_000019
|
| 41 |
+
berlin berlin_000079_000019
|
| 42 |
+
berlin berlin_000080_000019
|
| 43 |
+
berlin berlin_000081_000019
|
| 44 |
+
berlin berlin_000082_000019
|
| 45 |
+
berlin berlin_000083_000019
|
| 46 |
+
berlin berlin_000085_000019
|
| 47 |
+
berlin berlin_000086_000019
|
| 48 |
+
berlin berlin_000087_000019
|
| 49 |
+
berlin berlin_000088_000019
|
| 50 |
+
berlin berlin_000089_000019
|
| 51 |
+
berlin berlin_000090_000019
|
| 52 |
+
berlin berlin_000091_000019
|
| 53 |
+
berlin berlin_000092_000019
|
| 54 |
+
berlin berlin_000093_000019
|
| 55 |
+
berlin berlin_000094_000019
|
| 56 |
+
berlin berlin_000096_000019
|
| 57 |
+
berlin berlin_000097_000019
|
| 58 |
+
berlin berlin_000098_000019
|
| 59 |
+
berlin berlin_000099_000019
|
| 60 |
+
berlin berlin_000102_000019
|
| 61 |
+
berlin berlin_000106_000019
|
| 62 |
+
berlin berlin_000108_000019
|
| 63 |
+
berlin berlin_000109_000019
|
| 64 |
+
berlin berlin_000114_000019
|
| 65 |
+
berlin berlin_000116_000019
|
| 66 |
+
berlin berlin_000117_000019
|
| 67 |
+
berlin berlin_000122_000019
|
| 68 |
+
berlin berlin_000123_000019
|
| 69 |
+
berlin berlin_000126_000019
|
| 70 |
+
berlin berlin_000128_000019
|
| 71 |
+
berlin berlin_000131_000019
|
| 72 |
+
berlin berlin_000135_000019
|
| 73 |
+
berlin berlin_000137_000019
|
| 74 |
+
berlin berlin_000138_000019
|
| 75 |
+
berlin berlin_000141_000019
|
| 76 |
+
berlin berlin_000144_000019
|
| 77 |
+
berlin berlin_000145_000019
|
| 78 |
+
berlin berlin_000146_000019
|
| 79 |
+
berlin berlin_000147_000019
|
| 80 |
+
berlin berlin_000148_000019
|
| 81 |
+
berlin berlin_000149_000019
|
| 82 |
+
berlin berlin_000153_000019
|
| 83 |
+
berlin berlin_000155_000019
|
| 84 |
+
berlin berlin_000156_000019
|
| 85 |
+
berlin berlin_000157_000019
|
| 86 |
+
berlin berlin_000158_000019
|
| 87 |
+
berlin berlin_000162_000019
|
| 88 |
+
berlin berlin_000163_000019
|
| 89 |
+
berlin berlin_000164_000019
|
| 90 |
+
berlin berlin_000169_000019
|
| 91 |
+
berlin berlin_000171_000019
|
| 92 |
+
berlin berlin_000172_000019
|
| 93 |
+
berlin berlin_000178_000019
|
| 94 |
+
berlin berlin_000181_000019
|
| 95 |
+
berlin berlin_000182_000019
|
| 96 |
+
berlin berlin_000184_000019
|
| 97 |
+
berlin berlin_000185_000019
|
| 98 |
+
berlin berlin_000186_000019
|
| 99 |
+
berlin berlin_000187_000019
|
| 100 |
+
berlin berlin_000188_000019
|
| 101 |
+
berlin berlin_000192_000019
|
| 102 |
+
berlin berlin_000195_000019
|
| 103 |
+
berlin berlin_000196_000019
|
| 104 |
+
berlin berlin_000198_000019
|
| 105 |
+
berlin berlin_000203_000019
|
| 106 |
+
berlin berlin_000206_000019
|
| 107 |
+
berlin berlin_000211_000019
|
| 108 |
+
berlin berlin_000212_000019
|
| 109 |
+
berlin berlin_000215_000019
|
| 110 |
+
berlin berlin_000217_000019
|
| 111 |
+
berlin berlin_000222_000019
|
| 112 |
+
berlin berlin_000224_000019
|
| 113 |
+
berlin berlin_000226_000019
|
| 114 |
+
berlin berlin_000227_000019
|
| 115 |
+
berlin berlin_000229_000019
|
| 116 |
+
berlin berlin_000230_000019
|
| 117 |
+
berlin berlin_000233_000019
|
| 118 |
+
berlin berlin_000235_000019
|
| 119 |
+
berlin berlin_000236_000019
|
| 120 |
+
berlin berlin_000238_000019
|
| 121 |
+
berlin berlin_000239_000019
|
| 122 |
+
berlin berlin_000241_000019
|
| 123 |
+
berlin berlin_000244_000019
|
| 124 |
+
berlin berlin_000247_000019
|
| 125 |
+
berlin berlin_000248_000019
|
| 126 |
+
berlin berlin_000249_000019
|
| 127 |
+
berlin berlin_000255_000019
|
| 128 |
+
berlin berlin_000258_000019
|
| 129 |
+
berlin berlin_000262_000019
|
| 130 |
+
berlin berlin_000270_000019
|
| 131 |
+
berlin berlin_000271_000019
|
| 132 |
+
berlin berlin_000272_000019
|
| 133 |
+
berlin berlin_000275_000019
|
| 134 |
+
berlin berlin_000277_000019
|
| 135 |
+
berlin berlin_000278_000019
|
| 136 |
+
berlin berlin_000279_000019
|
| 137 |
+
berlin berlin_000280_000019
|
| 138 |
+
berlin berlin_000283_000019
|
| 139 |
+
berlin berlin_000286_000019
|
| 140 |
+
berlin berlin_000287_000019
|
| 141 |
+
berlin berlin_000292_000019
|
| 142 |
+
berlin berlin_000294_000019
|
| 143 |
+
berlin berlin_000299_000019
|
| 144 |
+
berlin berlin_000302_000019
|
| 145 |
+
berlin berlin_000305_000019
|
| 146 |
+
berlin berlin_000306_000019
|
| 147 |
+
berlin berlin_000311_000019
|
| 148 |
+
berlin berlin_000312_000019
|
| 149 |
+
berlin berlin_000318_000019
|
| 150 |
+
berlin berlin_000319_000019
|
| 151 |
+
berlin berlin_000326_000019
|
| 152 |
+
berlin berlin_000327_000019
|
| 153 |
+
berlin berlin_000328_000019
|
| 154 |
+
berlin berlin_000329_000019
|
| 155 |
+
berlin berlin_000331_000019
|
| 156 |
+
berlin berlin_000332_000019
|
| 157 |
+
berlin berlin_000333_000019
|
| 158 |
+
berlin berlin_000335_000019
|
| 159 |
+
berlin berlin_000338_000019
|
| 160 |
+
berlin berlin_000339_000019
|
| 161 |
+
berlin berlin_000340_000019
|
| 162 |
+
berlin berlin_000341_000019
|
| 163 |
+
berlin berlin_000343_000019
|
| 164 |
+
berlin berlin_000344_000019
|
| 165 |
+
berlin berlin_000346_000019
|
| 166 |
+
berlin berlin_000347_000019
|
| 167 |
+
berlin berlin_000348_000019
|
| 168 |
+
berlin berlin_000349_000019
|
| 169 |
+
berlin berlin_000350_000019
|
| 170 |
+
berlin berlin_000353_000019
|
| 171 |
+
berlin berlin_000354_000019
|
| 172 |
+
berlin berlin_000355_000019
|
| 173 |
+
berlin berlin_000356_000019
|
| 174 |
+
berlin berlin_000361_000019
|
| 175 |
+
berlin berlin_000363_000019
|
| 176 |
+
berlin berlin_000365_000019
|
| 177 |
+
berlin berlin_000366_000019
|
| 178 |
+
berlin berlin_000368_000019
|
| 179 |
+
berlin berlin_000371_000019
|
| 180 |
+
berlin berlin_000372_000019
|
| 181 |
+
berlin berlin_000374_000019
|
| 182 |
+
berlin berlin_000375_000019
|
| 183 |
+
berlin berlin_000376_000019
|
| 184 |
+
berlin berlin_000377_000019
|
| 185 |
+
berlin berlin_000380_000019
|
| 186 |
+
berlin berlin_000381_000019
|
| 187 |
+
berlin berlin_000382_000019
|
| 188 |
+
berlin berlin_000385_000019
|
| 189 |
+
berlin berlin_000387_000019
|
| 190 |
+
berlin berlin_000389_000019
|
| 191 |
+
berlin berlin_000390_000019
|
| 192 |
+
berlin berlin_000396_000019
|
| 193 |
+
berlin berlin_000397_000019
|
| 194 |
+
berlin berlin_000400_000019
|
| 195 |
+
berlin berlin_000401_000019
|
| 196 |
+
berlin berlin_000403_000019
|
| 197 |
+
berlin berlin_000404_000019
|
| 198 |
+
berlin berlin_000407_000019
|
| 199 |
+
berlin berlin_000409_000019
|
| 200 |
+
berlin berlin_000410_000019
|
| 201 |
+
berlin berlin_000412_000019
|
| 202 |
+
berlin berlin_000415_000019
|
| 203 |
+
berlin berlin_000416_000019
|
| 204 |
+
berlin berlin_000417_000019
|
| 205 |
+
berlin berlin_000418_000019
|
| 206 |
+
berlin berlin_000421_000019
|
| 207 |
+
berlin berlin_000422_000019
|
| 208 |
+
berlin berlin_000424_000019
|
| 209 |
+
berlin berlin_000425_000019
|
| 210 |
+
berlin berlin_000427_000019
|
| 211 |
+
berlin berlin_000429_000019
|
| 212 |
+
berlin berlin_000430_000019
|
| 213 |
+
berlin berlin_000431_000019
|
| 214 |
+
berlin berlin_000432_000019
|
| 215 |
+
berlin berlin_000435_000019
|
| 216 |
+
berlin berlin_000436_000019
|
| 217 |
+
berlin berlin_000437_000019
|
| 218 |
+
berlin berlin_000440_000019
|
| 219 |
+
berlin berlin_000444_000019
|
| 220 |
+
berlin berlin_000445_000019
|
| 221 |
+
berlin berlin_000447_000019
|
| 222 |
+
berlin berlin_000451_000019
|
| 223 |
+
berlin berlin_000453_000019
|
| 224 |
+
berlin berlin_000454_000019
|
| 225 |
+
berlin berlin_000456_000019
|
| 226 |
+
berlin berlin_000458_000019
|
| 227 |
+
berlin berlin_000461_000019
|
| 228 |
+
berlin berlin_000465_000019
|
| 229 |
+
berlin berlin_000466_000019
|
| 230 |
+
berlin berlin_000468_000019
|
| 231 |
+
berlin berlin_000469_000019
|
| 232 |
+
berlin berlin_000471_000019
|
| 233 |
+
berlin berlin_000472_000019
|
| 234 |
+
berlin berlin_000473_000019
|
| 235 |
+
berlin berlin_000479_000019
|
| 236 |
+
berlin berlin_000482_000019
|
| 237 |
+
berlin berlin_000484_000019
|
| 238 |
+
berlin berlin_000485_000019
|
| 239 |
+
berlin berlin_000486_000019
|
| 240 |
+
berlin berlin_000490_000019
|
| 241 |
+
berlin berlin_000491_000019
|
| 242 |
+
berlin berlin_000492_000019
|
| 243 |
+
berlin berlin_000493_000019
|
| 244 |
+
berlin berlin_000498_000019
|
| 245 |
+
berlin berlin_000502_000019
|
| 246 |
+
berlin berlin_000503_000019
|
| 247 |
+
berlin berlin_000504_000019
|
| 248 |
+
berlin berlin_000505_000019
|
| 249 |
+
berlin berlin_000507_000019
|
| 250 |
+
berlin berlin_000510_000019
|
| 251 |
+
berlin berlin_000511_000019
|
| 252 |
+
berlin berlin_000512_000019
|
| 253 |
+
berlin berlin_000514_000019
|
| 254 |
+
berlin berlin_000515_000019
|
| 255 |
+
berlin berlin_000516_000019
|
| 256 |
+
berlin berlin_000518_000019
|
| 257 |
+
berlin berlin_000519_000019
|
| 258 |
+
berlin berlin_000521_000019
|
| 259 |
+
berlin berlin_000522_000019
|
| 260 |
+
berlin berlin_000523_000019
|
| 261 |
+
berlin berlin_000524_000019
|
| 262 |
+
berlin berlin_000525_000019
|
| 263 |
+
berlin berlin_000527_000019
|
| 264 |
+
berlin berlin_000529_000019
|
| 265 |
+
berlin berlin_000530_000019
|
| 266 |
+
berlin berlin_000531_000019
|
| 267 |
+
berlin berlin_000532_000019
|
| 268 |
+
berlin berlin_000536_000019
|
| 269 |
+
berlin berlin_000539_000019
|
| 270 |
+
berlin berlin_000541_000019
|
| 271 |
+
bielefeld bielefeld_000000_001187
|
| 272 |
+
bielefeld bielefeld_000000_001505
|
| 273 |
+
bielefeld bielefeld_000000_002308
|
| 274 |
+
bielefeld bielefeld_000000_003406
|
| 275 |
+
bielefeld bielefeld_000000_003546
|
| 276 |
+
bielefeld bielefeld_000000_003731
|
| 277 |
+
bielefeld bielefeld_000000_005068
|
| 278 |
+
bielefeld bielefeld_000000_005584
|
| 279 |
+
bielefeld bielefeld_000000_006239
|
| 280 |
+
bielefeld bielefeld_000000_006802
|
| 281 |
+
bielefeld bielefeld_000000_007030
|
| 282 |
+
bielefeld bielefeld_000000_008279
|
| 283 |
+
bielefeld bielefeld_000000_008581
|
| 284 |
+
bielefeld bielefeld_000000_008800
|
| 285 |
+
bielefeld bielefeld_000000_009728
|
| 286 |
+
bielefeld bielefeld_000000_009928
|
| 287 |
+
bielefeld bielefeld_000000_011367
|
| 288 |
+
bielefeld bielefeld_000000_011831
|
| 289 |
+
bielefeld bielefeld_000000_012080
|
| 290 |
+
bielefeld bielefeld_000000_012788
|
| 291 |
+
bielefeld bielefeld_000000_013570
|
| 292 |
+
bielefeld bielefeld_000000_013814
|
| 293 |
+
bielefeld bielefeld_000000_015301
|
| 294 |
+
bielefeld bielefeld_000000_015867
|
| 295 |
+
bielefeld bielefeld_000000_016019
|
| 296 |
+
bielefeld bielefeld_000000_017051
|
| 297 |
+
bielefeld bielefeld_000000_017774
|
| 298 |
+
bielefeld bielefeld_000000_018102
|
| 299 |
+
bielefeld bielefeld_000000_018644
|
| 300 |
+
bielefeld bielefeld_000000_019416
|
| 301 |
+
bielefeld bielefeld_000000_020757
|
| 302 |
+
bielefeld bielefeld_000000_020900
|
| 303 |
+
bielefeld bielefeld_000000_021341
|
| 304 |
+
bielefeld bielefeld_000000_021381
|
| 305 |
+
bielefeld bielefeld_000000_022261
|
| 306 |
+
bielefeld bielefeld_000000_022835
|
| 307 |
+
bielefeld bielefeld_000000_025061
|
| 308 |
+
bielefeld bielefeld_000000_025748
|
| 309 |
+
bielefeld bielefeld_000000_026053
|
| 310 |
+
bielefeld bielefeld_000000_026550
|
| 311 |
+
bielefeld bielefeld_000000_026660
|
| 312 |
+
bielefeld bielefeld_000000_026823
|
| 313 |
+
bielefeld bielefeld_000000_027221
|
| 314 |
+
bielefeld bielefeld_000000_027586
|
| 315 |
+
bielefeld bielefeld_000000_027928
|
| 316 |
+
bielefeld bielefeld_000000_028148
|
| 317 |
+
bielefeld bielefeld_000000_028550
|
| 318 |
+
bielefeld bielefeld_000000_030038
|
| 319 |
+
bielefeld bielefeld_000000_030958
|
| 320 |
+
bielefeld bielefeld_000000_031244
|
| 321 |
+
bielefeld bielefeld_000000_032388
|
| 322 |
+
bielefeld bielefeld_000000_033770
|
| 323 |
+
bielefeld bielefeld_000000_035537
|
| 324 |
+
bielefeld bielefeld_000000_035879
|
| 325 |
+
bielefeld bielefeld_000000_036362
|
| 326 |
+
bielefeld bielefeld_000000_036732
|
| 327 |
+
bielefeld bielefeld_000000_037016
|
| 328 |
+
bielefeld bielefeld_000000_037422
|
| 329 |
+
bielefeld bielefeld_000000_038924
|
| 330 |
+
bielefeld bielefeld_000000_039221
|
| 331 |
+
bielefeld bielefeld_000000_040035
|
| 332 |
+
bielefeld bielefeld_000000_040472
|
| 333 |
+
bielefeld bielefeld_000000_041014
|
| 334 |
+
bielefeld bielefeld_000000_041142
|
| 335 |
+
bielefeld bielefeld_000000_041444
|
| 336 |
+
bielefeld bielefeld_000000_042403
|
| 337 |
+
bielefeld bielefeld_000000_042571
|
| 338 |
+
bielefeld bielefeld_000000_042717
|
| 339 |
+
bielefeld bielefeld_000000_043100
|
| 340 |
+
bielefeld bielefeld_000000_043389
|
| 341 |
+
bielefeld bielefeld_000000_043610
|
| 342 |
+
bielefeld bielefeld_000000_044085
|
| 343 |
+
bielefeld bielefeld_000000_045117
|
| 344 |
+
bielefeld bielefeld_000000_045232
|
| 345 |
+
bielefeld bielefeld_000000_046023
|
| 346 |
+
bielefeld bielefeld_000000_046212
|
| 347 |
+
bielefeld bielefeld_000000_046495
|
| 348 |
+
bielefeld bielefeld_000000_047918
|
| 349 |
+
bielefeld bielefeld_000000_048227
|
| 350 |
+
bielefeld bielefeld_000000_048518
|
| 351 |
+
bielefeld bielefeld_000000_049446
|
| 352 |
+
bielefeld bielefeld_000000_050586
|
| 353 |
+
bielefeld bielefeld_000000_051894
|
| 354 |
+
bielefeld bielefeld_000000_052155
|
| 355 |
+
bielefeld bielefeld_000000_053384
|
| 356 |
+
bielefeld bielefeld_000000_053583
|
| 357 |
+
bielefeld bielefeld_000000_055145
|
| 358 |
+
bielefeld bielefeld_000000_056175
|
| 359 |
+
bielefeld bielefeld_000000_056226
|
| 360 |
+
bielefeld bielefeld_000000_056603
|
| 361 |
+
bielefeld bielefeld_000000_056866
|
| 362 |
+
bielefeld bielefeld_000000_058374
|
| 363 |
+
bielefeld bielefeld_000000_058934
|
| 364 |
+
bielefeld bielefeld_000000_059303
|
| 365 |
+
bielefeld bielefeld_000000_059766
|
| 366 |
+
bielefeld bielefeld_000000_059842
|
| 367 |
+
bielefeld bielefeld_000000_060786
|
| 368 |
+
bielefeld bielefeld_000000_061094
|
| 369 |
+
bielefeld bielefeld_000000_062121
|
| 370 |
+
bielefeld bielefeld_000000_063623
|
| 371 |
+
bielefeld bielefeld_000000_063939
|
| 372 |
+
bielefeld bielefeld_000000_064271
|
| 373 |
+
bielefeld bielefeld_000000_064805
|
| 374 |
+
bielefeld bielefeld_000000_064910
|
| 375 |
+
bielefeld bielefeld_000000_066195
|
| 376 |
+
bielefeld bielefeld_000000_066405
|
| 377 |
+
bielefeld bielefeld_000000_066495
|
| 378 |
+
bonn bonn_000001_000019
|
| 379 |
+
bonn bonn_000002_000019
|
| 380 |
+
bonn bonn_000005_000019
|
| 381 |
+
bonn bonn_000006_000019
|
| 382 |
+
bonn bonn_000010_000019
|
| 383 |
+
bonn bonn_000013_000019
|
| 384 |
+
bonn bonn_000016_000019
|
| 385 |
+
bonn bonn_000020_000019
|
| 386 |
+
bonn bonn_000021_000019
|
| 387 |
+
bonn bonn_000022_000019
|
| 388 |
+
bonn bonn_000026_000019
|
| 389 |
+
bonn bonn_000027_000019
|
| 390 |
+
bonn bonn_000028_000019
|
| 391 |
+
bonn bonn_000029_000019
|
| 392 |
+
bonn bonn_000031_000019
|
| 393 |
+
bonn bonn_000033_000019
|
| 394 |
+
bonn bonn_000034_000019
|
| 395 |
+
bonn bonn_000037_000019
|
| 396 |
+
bonn bonn_000038_000019
|
| 397 |
+
bonn bonn_000039_000019
|
| 398 |
+
bonn bonn_000040_000019
|
| 399 |
+
bonn bonn_000041_000019
|
| 400 |
+
bonn bonn_000042_000019
|
| 401 |
+
bonn bonn_000043_000019
|
| 402 |
+
bonn bonn_000044_000019
|
| 403 |
+
leverkusen leverkusen_000003_000019
|
| 404 |
+
leverkusen leverkusen_000005_000019
|
| 405 |
+
leverkusen leverkusen_000006_000019
|
| 406 |
+
leverkusen leverkusen_000009_000019
|
| 407 |
+
leverkusen leverkusen_000010_000019
|
| 408 |
+
leverkusen leverkusen_000011_000019
|
| 409 |
+
leverkusen leverkusen_000012_000019
|
| 410 |
+
leverkusen leverkusen_000017_000019
|
| 411 |
+
leverkusen leverkusen_000018_000019
|
| 412 |
+
leverkusen leverkusen_000020_000019
|
| 413 |
+
leverkusen leverkusen_000021_000019
|
| 414 |
+
leverkusen leverkusen_000022_000019
|
| 415 |
+
leverkusen leverkusen_000023_000019
|
| 416 |
+
leverkusen leverkusen_000027_000019
|
| 417 |
+
leverkusen leverkusen_000029_000019
|
| 418 |
+
leverkusen leverkusen_000030_000019
|
| 419 |
+
leverkusen leverkusen_000031_000019
|
| 420 |
+
leverkusen leverkusen_000032_000019
|
| 421 |
+
leverkusen leverkusen_000034_000019
|
| 422 |
+
leverkusen leverkusen_000035_000019
|
| 423 |
+
leverkusen leverkusen_000036_000019
|
| 424 |
+
leverkusen leverkusen_000040_000019
|
| 425 |
+
leverkusen leverkusen_000041_000019
|
| 426 |
+
leverkusen leverkusen_000042_000019
|
| 427 |
+
leverkusen leverkusen_000044_000019
|
| 428 |
+
leverkusen leverkusen_000045_000019
|
| 429 |
+
leverkusen leverkusen_000048_000019
|
| 430 |
+
leverkusen leverkusen_000049_000019
|
| 431 |
+
leverkusen leverkusen_000050_000019
|
| 432 |
+
leverkusen leverkusen_000051_000019
|
| 433 |
+
leverkusen leverkusen_000052_000019
|
| 434 |
+
leverkusen leverkusen_000054_000019
|
| 435 |
+
leverkusen leverkusen_000056_000019
|
| 436 |
+
mainz mainz_000000_000093
|
| 437 |
+
mainz mainz_000000_000293
|
| 438 |
+
mainz mainz_000000_001003
|
| 439 |
+
mainz mainz_000000_001068
|
| 440 |
+
mainz mainz_000000_001265
|
| 441 |
+
mainz mainz_000000_001410
|
| 442 |
+
mainz mainz_000000_001857
|
| 443 |
+
mainz mainz_000000_003250
|
| 444 |
+
mainz mainz_000000_007415
|
| 445 |
+
mainz mainz_000000_008165
|
| 446 |
+
mainz mainz_000000_008645
|
| 447 |
+
mainz mainz_000000_008871
|
| 448 |
+
mainz mainz_000000_009985
|
| 449 |
+
mainz mainz_000000_010550
|
| 450 |
+
mainz mainz_000000_011339
|
| 451 |
+
mainz mainz_000000_011879
|
| 452 |
+
mainz mainz_000000_011965
|
| 453 |
+
mainz mainz_000000_012392
|
| 454 |
+
mainz mainz_000000_013095
|
| 455 |
+
mainz mainz_000000_013437
|
| 456 |
+
mainz mainz_000000_013960
|
| 457 |
+
mainz mainz_000000_014193
|
| 458 |
+
mainz mainz_000000_014742
|
| 459 |
+
mainz mainz_000000_015052
|
| 460 |
+
mainz mainz_000000_015170
|
| 461 |
+
mainz mainz_000000_015760
|
| 462 |
+
mainz mainz_000000_017927
|
| 463 |
+
mainz mainz_000000_018249
|
| 464 |
+
mainz mainz_000000_018883
|
| 465 |
+
mainz mainz_000000_019043
|
| 466 |
+
mainz mainz_000000_019227
|
| 467 |
+
mainz mainz_000000_019847
|
| 468 |
+
mainz mainz_000000_020415
|
| 469 |
+
mainz mainz_000000_021457
|
| 470 |
+
mainz mainz_000000_021524
|
| 471 |
+
mainz mainz_000000_021735
|
| 472 |
+
mainz mainz_000000_022417
|
| 473 |
+
mainz mainz_000001_000120
|
| 474 |
+
mainz mainz_000001_000428
|
| 475 |
+
mainz mainz_000001_001509
|
| 476 |
+
mainz mainz_000001_002884
|
| 477 |
+
mainz mainz_000001_003012
|
| 478 |
+
mainz mainz_000001_003624
|
| 479 |
+
mainz mainz_000001_004132
|
| 480 |
+
mainz mainz_000001_004219
|
| 481 |
+
mainz mainz_000001_004823
|
| 482 |
+
mainz mainz_000001_005016
|
| 483 |
+
mainz mainz_000001_005163
|
| 484 |
+
mainz mainz_000001_005665
|
| 485 |
+
mainz mainz_000001_005815
|
| 486 |
+
mainz mainz_000001_005911
|
| 487 |
+
mainz mainz_000001_006194
|
| 488 |
+
mainz mainz_000001_007595
|
| 489 |
+
mainz mainz_000001_007956
|
| 490 |
+
mainz mainz_000001_008056
|
| 491 |
+
mainz mainz_000001_008264
|
| 492 |
+
mainz mainz_000001_008540
|
| 493 |
+
mainz mainz_000001_008638
|
| 494 |
+
mainz mainz_000001_009328
|
| 495 |
+
mainz mainz_000001_009867
|
| 496 |
+
mainz mainz_000001_009996
|
| 497 |
+
mainz mainz_000001_010853
|
| 498 |
+
mainz mainz_000001_011736
|
| 499 |
+
mainz mainz_000001_012541
|
| 500 |
+
mainz mainz_000001_014073
|
| 501 |
+
mainz mainz_000001_014469
|
| 502 |
+
mainz mainz_000001_015117
|
| 503 |
+
mainz mainz_000001_015508
|
| 504 |
+
mainz mainz_000001_015724
|
| 505 |
+
mainz mainz_000001_018145
|
| 506 |
+
mainz mainz_000001_019061
|
| 507 |
+
mainz mainz_000001_019593
|
| 508 |
+
mainz mainz_000001_020068
|
| 509 |
+
mainz mainz_000001_020484
|
| 510 |
+
mainz mainz_000001_020829
|
| 511 |
+
mainz mainz_000001_022125
|
| 512 |
+
mainz mainz_000001_022630
|
| 513 |
+
mainz mainz_000001_023604
|
| 514 |
+
mainz mainz_000001_024489
|
| 515 |
+
mainz mainz_000001_024718
|
| 516 |
+
mainz mainz_000001_025623
|
| 517 |
+
mainz mainz_000001_026837
|
| 518 |
+
mainz mainz_000001_027053
|
| 519 |
+
mainz mainz_000001_027377
|
| 520 |
+
mainz mainz_000001_027751
|
| 521 |
+
mainz mainz_000001_028326
|
| 522 |
+
mainz mainz_000001_028566
|
| 523 |
+
mainz mainz_000001_029755
|
| 524 |
+
mainz mainz_000001_029950
|
| 525 |
+
mainz mainz_000001_030417
|
| 526 |
+
mainz mainz_000001_031697
|
| 527 |
+
mainz mainz_000001_031946
|
| 528 |
+
mainz mainz_000001_032294
|
| 529 |
+
mainz mainz_000001_032401
|
| 530 |
+
mainz mainz_000001_032691
|
| 531 |
+
mainz mainz_000001_032767
|
| 532 |
+
mainz mainz_000001_032911
|
| 533 |
+
mainz mainz_000001_033096
|
| 534 |
+
mainz mainz_000001_033603
|
| 535 |
+
mainz mainz_000001_034209
|
| 536 |
+
mainz mainz_000001_034681
|
| 537 |
+
mainz mainz_000001_035293
|
| 538 |
+
mainz mainz_000001_035963
|
| 539 |
+
mainz mainz_000001_036240
|
| 540 |
+
mainz mainz_000001_036412
|
| 541 |
+
mainz mainz_000001_037905
|
| 542 |
+
mainz mainz_000001_038026
|
| 543 |
+
mainz mainz_000001_038191
|
| 544 |
+
mainz mainz_000001_038768
|
| 545 |
+
mainz mainz_000001_039075
|
| 546 |
+
mainz mainz_000001_039470
|
| 547 |
+
mainz mainz_000001_040367
|
| 548 |
+
mainz mainz_000001_041172
|
| 549 |
+
mainz mainz_000001_041647
|
| 550 |
+
mainz mainz_000001_041887
|
| 551 |
+
mainz mainz_000001_042121
|
| 552 |
+
mainz mainz_000001_042851
|
| 553 |
+
mainz mainz_000001_043656
|
| 554 |
+
mainz mainz_000001_044366
|
| 555 |
+
mainz mainz_000001_045385
|
| 556 |
+
mainz mainz_000001_045651
|
| 557 |
+
mainz mainz_000001_045782
|
| 558 |
+
mainz mainz_000001_046381
|
| 559 |
+
mainz mainz_000001_047546
|
| 560 |
+
mainz mainz_000001_047611
|
| 561 |
+
mainz mainz_000001_047888
|
| 562 |
+
mainz mainz_000001_048725
|
| 563 |
+
mainz mainz_000002_000061
|
| 564 |
+
mainz mainz_000002_000912
|
| 565 |
+
mainz mainz_000003_001465
|
| 566 |
+
mainz mainz_000003_001899
|
| 567 |
+
mainz mainz_000003_003558
|
| 568 |
+
mainz mainz_000003_003711
|
| 569 |
+
mainz mainz_000003_004144
|
| 570 |
+
mainz mainz_000003_004228
|
| 571 |
+
mainz mainz_000003_004774
|
| 572 |
+
mainz mainz_000003_005088
|
| 573 |
+
mainz mainz_000003_006863
|
| 574 |
+
mainz mainz_000003_007144
|
| 575 |
+
mainz mainz_000003_008690
|
| 576 |
+
mainz mainz_000003_008876
|
| 577 |
+
mainz mainz_000003_009819
|
| 578 |
+
mainz mainz_000003_010772
|
| 579 |
+
mainz mainz_000003_010924
|
| 580 |
+
mainz mainz_000003_011949
|
| 581 |
+
mainz mainz_000003_012341
|
| 582 |
+
mainz mainz_000003_012995
|
| 583 |
+
mainz mainz_000003_014083
|
| 584 |
+
mainz mainz_000003_014457
|
| 585 |
+
mainz mainz_000003_014537
|
| 586 |
+
mainz mainz_000003_014959
|
| 587 |
+
mainz mainz_000003_015411
|
| 588 |
+
mainz mainz_000003_015649
|
| 589 |
+
mainz mainz_000003_015917
|
| 590 |
+
mainz mainz_000003_016360
|
| 591 |
+
mainz mainz_000003_016542
|
| 592 |
+
mainz mainz_000003_017171
|
| 593 |
+
munich munich_000000_000019
|
| 594 |
+
munich munich_000002_000019
|
| 595 |
+
munich munich_000003_000019
|
| 596 |
+
munich munich_000005_000019
|
| 597 |
+
munich munich_000006_000019
|
| 598 |
+
munich munich_000009_000019
|
| 599 |
+
munich munich_000011_000019
|
| 600 |
+
munich munich_000012_000019
|
| 601 |
+
munich munich_000016_000019
|
| 602 |
+
munich munich_000017_000019
|
| 603 |
+
munich munich_000019_000019
|
| 604 |
+
munich munich_000025_000019
|
| 605 |
+
munich munich_000026_000019
|
| 606 |
+
munich munich_000028_000019
|
| 607 |
+
munich munich_000030_000019
|
| 608 |
+
munich munich_000032_000019
|
| 609 |
+
munich munich_000038_000019
|
| 610 |
+
munich munich_000040_000019
|
| 611 |
+
munich munich_000042_000019
|
| 612 |
+
munich munich_000046_000019
|
| 613 |
+
munich munich_000047_000019
|
| 614 |
+
munich munich_000050_000019
|
| 615 |
+
munich munich_000051_000019
|
| 616 |
+
munich munich_000052_000019
|
| 617 |
+
munich munich_000054_000019
|
| 618 |
+
munich munich_000055_000019
|
| 619 |
+
munich munich_000056_000019
|
| 620 |
+
munich munich_000057_000019
|
| 621 |
+
munich munich_000058_000019
|
| 622 |
+
munich munich_000062_000019
|
| 623 |
+
munich munich_000063_000019
|
| 624 |
+
munich munich_000065_000019
|
| 625 |
+
munich munich_000066_000019
|
| 626 |
+
munich munich_000067_000019
|
| 627 |
+
munich munich_000068_000019
|
| 628 |
+
munich munich_000070_000019
|
| 629 |
+
munich munich_000072_000019
|
| 630 |
+
munich munich_000075_000019
|
| 631 |
+
munich munich_000077_000019
|
| 632 |
+
munich munich_000079_000019
|
| 633 |
+
munich munich_000081_000019
|
| 634 |
+
munich munich_000082_000019
|
| 635 |
+
munich munich_000083_000019
|
| 636 |
+
munich munich_000084_000019
|
| 637 |
+
munich munich_000085_000019
|
| 638 |
+
munich munich_000088_000019
|
| 639 |
+
munich munich_000090_000019
|
| 640 |
+
munich munich_000091_000019
|
| 641 |
+
munich munich_000093_000019
|
| 642 |
+
munich munich_000096_000019
|
| 643 |
+
munich munich_000097_000019
|
| 644 |
+
munich munich_000098_000019
|
| 645 |
+
munich munich_000100_000019
|
| 646 |
+
munich munich_000101_000019
|
| 647 |
+
munich munich_000103_000019
|
| 648 |
+
munich munich_000109_000019
|
| 649 |
+
munich munich_000110_000019
|
| 650 |
+
munich munich_000111_000019
|
| 651 |
+
munich munich_000113_000019
|
| 652 |
+
munich munich_000116_000019
|
| 653 |
+
munich munich_000118_000019
|
| 654 |
+
munich munich_000120_000019
|
| 655 |
+
munich munich_000121_000019
|
| 656 |
+
munich munich_000125_000019
|
| 657 |
+
munich munich_000126_000019
|
| 658 |
+
munich munich_000128_000019
|
| 659 |
+
munich munich_000131_000019
|
| 660 |
+
munich munich_000132_000019
|
| 661 |
+
munich munich_000134_000019
|
| 662 |
+
munich munich_000135_000019
|
| 663 |
+
munich munich_000137_000019
|
| 664 |
+
munich munich_000138_000019
|
| 665 |
+
munich munich_000139_000019
|
| 666 |
+
munich munich_000140_000019
|
| 667 |
+
munich munich_000141_000019
|
| 668 |
+
munich munich_000142_000019
|
| 669 |
+
munich munich_000143_000019
|
| 670 |
+
munich munich_000144_000019
|
| 671 |
+
munich munich_000145_000019
|
| 672 |
+
munich munich_000150_000019
|
| 673 |
+
munich munich_000154_000019
|
| 674 |
+
munich munich_000155_000019
|
| 675 |
+
munich munich_000156_000019
|
| 676 |
+
munich munich_000157_000019
|
| 677 |
+
munich munich_000161_000019
|
| 678 |
+
munich munich_000164_000019
|
| 679 |
+
munich munich_000166_000019
|
| 680 |
+
munich munich_000168_000019
|
| 681 |
+
munich munich_000170_000019
|
| 682 |
+
munich munich_000173_000019
|
| 683 |
+
munich munich_000174_000019
|
| 684 |
+
munich munich_000175_000019
|
| 685 |
+
munich munich_000176_000019
|
| 686 |
+
munich munich_000178_000019
|
| 687 |
+
munich munich_000179_000019
|
| 688 |
+
munich munich_000180_000019
|
| 689 |
+
munich munich_000183_000019
|
| 690 |
+
munich munich_000186_000019
|
| 691 |
+
munich munich_000187_000019
|
| 692 |
+
munich munich_000189_000019
|
| 693 |
+
munich munich_000193_000019
|
| 694 |
+
munich munich_000195_000019
|
| 695 |
+
munich munich_000197_000019
|
| 696 |
+
munich munich_000199_000019
|
| 697 |
+
munich munich_000200_000019
|
| 698 |
+
munich munich_000203_000019
|
| 699 |
+
munich munich_000204_000019
|
| 700 |
+
munich munich_000206_000019
|
| 701 |
+
munich munich_000207_000019
|
| 702 |
+
munich munich_000209_000019
|
| 703 |
+
munich munich_000210_000019
|
| 704 |
+
munich munich_000212_000019
|
| 705 |
+
munich munich_000215_000019
|
| 706 |
+
munich munich_000216_000019
|
| 707 |
+
munich munich_000217_000019
|
| 708 |
+
munich munich_000219_000019
|
| 709 |
+
munich munich_000220_000019
|
| 710 |
+
munich munich_000221_000019
|
| 711 |
+
munich munich_000223_000019
|
| 712 |
+
munich munich_000224_000019
|
| 713 |
+
munich munich_000225_000019
|
| 714 |
+
munich munich_000227_000019
|
| 715 |
+
munich munich_000230_000019
|
| 716 |
+
munich munich_000231_000019
|
| 717 |
+
munich munich_000234_000019
|
| 718 |
+
munich munich_000235_000019
|
| 719 |
+
munich munich_000237_000019
|
| 720 |
+
munich munich_000238_000019
|
| 721 |
+
munich munich_000241_000019
|
| 722 |
+
munich munich_000245_000019
|
| 723 |
+
munich munich_000246_000019
|
| 724 |
+
munich munich_000247_000019
|
| 725 |
+
munich munich_000251_000019
|
| 726 |
+
munich munich_000255_000019
|
| 727 |
+
munich munich_000256_000019
|
| 728 |
+
munich munich_000257_000019
|
| 729 |
+
munich munich_000258_000019
|
| 730 |
+
munich munich_000260_000019
|
| 731 |
+
munich munich_000261_000019
|
| 732 |
+
munich munich_000264_000019
|
| 733 |
+
munich munich_000265_000019
|
| 734 |
+
munich munich_000266_000019
|
| 735 |
+
munich munich_000270_000019
|
| 736 |
+
munich munich_000271_000019
|
| 737 |
+
munich munich_000272_000019
|
| 738 |
+
munich munich_000273_000019
|
| 739 |
+
munich munich_000275_000019
|
| 740 |
+
munich munich_000276_000019
|
| 741 |
+
munich munich_000280_000019
|
| 742 |
+
munich munich_000281_000019
|
| 743 |
+
munich munich_000282_000019
|
| 744 |
+
munich munich_000283_000019
|
| 745 |
+
munich munich_000285_000019
|
| 746 |
+
munich munich_000286_000019
|
| 747 |
+
munich munich_000287_000019
|
| 748 |
+
munich munich_000288_000019
|
| 749 |
+
munich munich_000289_000019
|
| 750 |
+
munich munich_000290_000019
|
| 751 |
+
munich munich_000295_000019
|
| 752 |
+
munich munich_000297_000019
|
| 753 |
+
munich munich_000298_000019
|
| 754 |
+
munich munich_000300_000019
|
| 755 |
+
munich munich_000303_000019
|
| 756 |
+
munich munich_000305_000019
|
| 757 |
+
munich munich_000307_000019
|
| 758 |
+
munich munich_000309_000019
|
| 759 |
+
munich munich_000312_000019
|
| 760 |
+
munich munich_000314_000019
|
| 761 |
+
munich munich_000318_000019
|
| 762 |
+
munich munich_000320_000019
|
| 763 |
+
munich munich_000322_000019
|
| 764 |
+
munich munich_000326_000019
|
| 765 |
+
munich munich_000328_000019
|
| 766 |
+
munich munich_000329_000019
|
| 767 |
+
munich munich_000331_000019
|
| 768 |
+
munich munich_000334_000019
|
| 769 |
+
munich munich_000335_000019
|
| 770 |
+
munich munich_000336_000019
|
| 771 |
+
munich munich_000338_000019
|
| 772 |
+
munich munich_000339_000019
|
| 773 |
+
munich munich_000343_000019
|
| 774 |
+
munich munich_000347_000019
|
| 775 |
+
munich munich_000348_000019
|
| 776 |
+
munich munich_000350_000019
|
| 777 |
+
munich munich_000351_000019
|
| 778 |
+
munich munich_000352_000019
|
| 779 |
+
munich munich_000354_000019
|
| 780 |
+
munich munich_000355_000019
|
| 781 |
+
munich munich_000356_000019
|
| 782 |
+
munich munich_000357_000019
|
| 783 |
+
munich munich_000358_000019
|
| 784 |
+
munich munich_000359_000019
|
| 785 |
+
munich munich_000361_000019
|
| 786 |
+
munich munich_000362_000019
|
| 787 |
+
munich munich_000366_000019
|
| 788 |
+
munich munich_000370_000019
|
| 789 |
+
munich munich_000373_000019
|
| 790 |
+
munich munich_000375_000019
|
| 791 |
+
munich munich_000386_000019
|
| 792 |
+
munich munich_000387_000019
|
| 793 |
+
munich munich_000388_000019
|
| 794 |
+
munich munich_000389_000019
|
| 795 |
+
munich munich_000391_000019
|
| 796 |
+
munich munich_000392_000019
|
| 797 |
+
munich munich_000393_000019
|
| 798 |
+
munich munich_000394_000019
|
| 799 |
+
munich munich_000395_000019
|
| 800 |
+
munich munich_000397_000019
|
external/WildCamera/splits/cityscapes_train.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|