Qnancy commited on
Commit
310b910
·
verified ·
1 Parent(s): cdf6bfb

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/__init__.py +0 -0
  2. external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/helpers.py +141 -0
  3. external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/imagebind_model.py +517 -0
  4. external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/multimodal_preprocessors.py +687 -0
  5. external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/transformer.py +284 -0
  6. external/Grounded-Segment-Anything/playground/PaintByExample/README.md +70 -0
  7. external/Grounded-Segment-Anything/playground/PaintByExample/paint_by_example.py +53 -0
  8. external/Grounded-Segment-Anything/voxelnext_3d_box/models/__init__.py +0 -0
  9. external/Grounded-Segment-Anything/voxelnext_3d_box/models/data_processor.py +211 -0
  10. external/Grounded-Segment-Anything/voxelnext_3d_box/models/mean_vfe.py +25 -0
  11. external/Grounded-Segment-Anything/voxelnext_3d_box/models/spconv_backbone_voxelnext.py +292 -0
  12. external/Grounded-Segment-Anything/voxelnext_3d_box/models/voxelnext_head.py +166 -0
  13. external/Grounded-Segment-Anything/voxelnext_3d_box/utils/centernet_utils.py +116 -0
  14. external/Grounded-Segment-Anything/voxelnext_3d_box/utils/config.py +34 -0
  15. external/Grounded-Segment-Anything/voxelnext_3d_box/utils/image_projection.py +103 -0
  16. external/WildCamera/WildCamera/benchmark/benchmark_calibration.py +98 -0
  17. external/WildCamera/WildCamera/benchmark/benchmark_crop.py +63 -0
  18. external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_megadepth.py +168 -0
  19. external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_scannet.py +165 -0
  20. external/WildCamera/WildCamera/datasets/GSV.py +73 -0
  21. external/WildCamera/WildCamera/datasets/GenericDataset.py +171 -0
  22. external/WildCamera/WildCamera/datasets/IncdDataset.py +110 -0
  23. external/WildCamera/WildCamera/datasets/MegaDepth.py +76 -0
  24. external/WildCamera/WildCamera/evaluation/evaluate_crop.py +157 -0
  25. external/WildCamera/WildCamera/evaluation/evaluate_fov.py +142 -0
  26. external/WildCamera/WildCamera/evaluation/evaluate_intrinsic.py +144 -0
  27. external/WildCamera/WildCamera/evaluation/evaluate_pose.py +209 -0
  28. external/WildCamera/WildCamera/newcrfs/__init__.py +0 -0
  29. external/WildCamera/WildCamera/newcrfs/__pycache__/swin_transformer.cpython-310.pyc +0 -0
  30. external/WildCamera/WildCamera/newcrfs/__pycache__/uper_crf_head.cpython-310.pyc +0 -0
  31. external/WildCamera/WildCamera/newcrfs/newcrf_incidencefield.py +196 -0
  32. external/WildCamera/WildCamera/newcrfs/newcrf_layers.py +433 -0
  33. external/WildCamera/WildCamera/newcrfs/newcrf_utils.py +264 -0
  34. external/WildCamera/WildCamera/newcrfs/swin_transformer.py +619 -0
  35. external/WildCamera/WildCamera/newcrfs/uper_crf_head.py +350 -0
  36. external/WildCamera/WildCamera/train/train_calibrator.py +297 -0
  37. external/WildCamera/__pycache__/hubconf.cpython-310.pyc +0 -0
  38. external/WildCamera/asset/download_demo_images.sh +8 -0
  39. external/WildCamera/asset/download_wildcamera_checkpoint.sh +13 -0
  40. external/WildCamera/asset/download_wildcamera_dataset.sh +0 -0
  41. external/WildCamera/demo/demo_dollyzoom.py +133 -0
  42. external/WildCamera/demo/demo_inference.py +30 -0
  43. external/WildCamera/demo/demo_restoration.py +33 -0
  44. external/WildCamera/splits/arkitscenes_test.txt +800 -0
  45. external/WildCamera/splits/arkitscenes_train.txt +0 -0
  46. external/WildCamera/splits/arkitscenes_val.txt +800 -0
  47. external/WildCamera/splits/biwirgbdid_test.txt +800 -0
  48. external/WildCamera/splits/cad120_test.txt +800 -0
  49. external/WildCamera/splits/cityscapes_test.txt +800 -0
  50. external/WildCamera/splits/cityscapes_train.txt +0 -0
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/__init__.py ADDED
File without changes
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/helpers.py ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ # Portions Copyright (c) Meta Platforms, Inc. and affiliates.
3
+ # All rights reserved.
4
+
5
+ # This source code is licensed under the license found in the
6
+ # LICENSE file in the root directory of this source tree.
7
+
8
+ import math
9
+
10
+ import einops
11
+ import numpy as np
12
+ import torch
13
+
14
+ import torch.nn as nn
15
+
16
+
17
+ class Normalize(nn.Module):
18
+ def __init__(self, dim: int) -> None:
19
+ super().__init__()
20
+ self.dim = dim
21
+
22
+ def forward(self, x):
23
+ return torch.nn.functional.normalize(x, dim=self.dim, p=2)
24
+
25
+
26
+ class LearnableLogitScaling(nn.Module):
27
+ def __init__(
28
+ self,
29
+ logit_scale_init: float = 1 / 0.07,
30
+ learnable: bool = True,
31
+ max_logit_scale: float = 100,
32
+ ) -> None:
33
+ super().__init__()
34
+ self.max_logit_scale = max_logit_scale
35
+ self.logit_scale_init = logit_scale_init
36
+ self.learnable = learnable
37
+ log_logit_scale = torch.ones([]) * np.log(self.logit_scale_init)
38
+ if learnable:
39
+ self.log_logit_scale = nn.Parameter(log_logit_scale)
40
+ else:
41
+ self.register_buffer("log_logit_scale", log_logit_scale)
42
+
43
+ def forward(self, x):
44
+ return torch.clip(self.log_logit_scale.exp(), max=self.max_logit_scale) * x
45
+
46
+ def extra_repr(self):
47
+ st = f"logit_scale_init={self.logit_scale_init},learnable={self.learnable}, max_logit_scale={self.max_logit_scale}"
48
+ return st
49
+
50
+
51
+ class EinOpsRearrange(nn.Module):
52
+ def __init__(self, rearrange_expr: str, **kwargs) -> None:
53
+ super().__init__()
54
+ self.rearrange_expr = rearrange_expr
55
+ self.kwargs = kwargs
56
+
57
+ def forward(self, x):
58
+ assert isinstance(x, torch.Tensor)
59
+ return einops.rearrange(x, self.rearrange_expr, **self.kwargs)
60
+
61
+
62
+ class VerboseNNModule(nn.Module):
63
+ """
64
+ Wrapper around nn.Module that prints registered buffers and parameter names.
65
+ """
66
+
67
+ @staticmethod
68
+ def get_readable_tensor_repr(name: str, tensor: torch.Tensor) -> str:
69
+ st = (
70
+ "("
71
+ + name
72
+ + "): "
73
+ + "tensor("
74
+ + str(tuple(tensor[1].shape))
75
+ + ", requires_grad="
76
+ + str(tensor[1].requires_grad)
77
+ + ")\n"
78
+ )
79
+ return st
80
+
81
+ def extra_repr(self) -> str:
82
+ named_modules = set()
83
+ for p in self.named_modules():
84
+ named_modules.update([p[0]])
85
+ named_modules = list(named_modules)
86
+
87
+ string_repr = ""
88
+ for p in self.named_parameters():
89
+ name = p[0].split(".")[0]
90
+ if name not in named_modules:
91
+ string_repr += self.get_readable_tensor_repr(name, p)
92
+
93
+ for p in self.named_buffers():
94
+ name = p[0].split(".")[0]
95
+ string_repr += self.get_readable_tensor_repr(name, p)
96
+
97
+ return string_repr
98
+
99
+
100
+ def cast_if_src_dtype(
101
+ tensor: torch.Tensor, src_dtype: torch.dtype, tgt_dtype: torch.dtype
102
+ ):
103
+ updated = False
104
+ if tensor.dtype == src_dtype:
105
+ tensor = tensor.to(dtype=tgt_dtype)
106
+ updated = True
107
+ return tensor, updated
108
+
109
+
110
+ class QuickGELU(nn.Module):
111
+ # From https://github.com/openai/CLIP/blob/d50d76daa670286dd6cacf3bcd80b5e4823fc8e1/clip/model.py#L166
112
+ def forward(self, x: torch.Tensor):
113
+ return x * torch.sigmoid(1.702 * x)
114
+
115
+
116
+ class SelectElement(nn.Module):
117
+ def __init__(self, index) -> None:
118
+ super().__init__()
119
+ self.index = index
120
+
121
+ def forward(self, x):
122
+ assert x.ndim >= 3
123
+ return x[:, self.index, ...]
124
+
125
+
126
+ class SelectEOSAndProject(nn.Module):
127
+ """
128
+ Text Pooling used in OpenCLIP
129
+ """
130
+
131
+ def __init__(self, proj: nn.Module) -> None:
132
+ super().__init__()
133
+ self.proj = proj
134
+
135
+ def forward(self, x, seq_len):
136
+ assert x.ndim == 3
137
+ # x is of shape B x L x D
138
+ # take features from the eot embedding (eot_token is the highest number in each sequence)
139
+ x = x[torch.arange(x.shape[0]), seq_len]
140
+ x = self.proj(x)
141
+ return x
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/imagebind_model.py ADDED
@@ -0,0 +1,517 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ # Portions Copyright (c) Meta Platforms, Inc. and affiliates.
3
+ # All rights reserved.
4
+
5
+ # This source code is licensed under the license found in the
6
+ # LICENSE file in the root directory of this source tree.
7
+
8
+
9
+ import os
10
+ import urllib
11
+ from functools import partial
12
+ from types import SimpleNamespace
13
+
14
+ import torch
15
+ import torch.nn as nn
16
+
17
+ from models.helpers import (
18
+ EinOpsRearrange,
19
+ LearnableLogitScaling,
20
+ Normalize,
21
+ SelectElement,
22
+ SelectEOSAndProject,
23
+ )
24
+ from models.multimodal_preprocessors import (
25
+ AudioPreprocessor,
26
+ IMUPreprocessor,
27
+ PadIm2Video,
28
+ PatchEmbedGeneric,
29
+ RGBDTPreprocessor,
30
+ SpatioTemporalPosEmbeddingHelper,
31
+ TextPreprocessor,
32
+ ThermalPreprocessor,
33
+ )
34
+
35
+ from models.transformer import MultiheadAttention, SimpleTransformer
36
+
37
+
38
+ ModalityType = SimpleNamespace(
39
+ VISION="vision",
40
+ TEXT="text",
41
+ AUDIO="audio",
42
+ THERMAL="thermal",
43
+ DEPTH="depth",
44
+ IMU="imu",
45
+ )
46
+
47
+
48
+ class ImageBindModel(nn.Module):
49
+ def __init__(
50
+ self,
51
+ video_frames=2,
52
+ kernel_size=(2, 14, 14),
53
+ audio_kernel_size=16,
54
+ audio_stride=10,
55
+ out_embed_dim=768,
56
+ vision_embed_dim=1024,
57
+ vision_num_blocks=24,
58
+ vision_num_heads=16,
59
+ audio_embed_dim=768,
60
+ audio_num_blocks=12,
61
+ audio_num_heads=12,
62
+ audio_num_mel_bins=128,
63
+ audio_target_len=204,
64
+ audio_drop_path=0.1,
65
+ text_embed_dim=768,
66
+ text_num_blocks=12,
67
+ text_num_heads=12,
68
+ depth_embed_dim=384,
69
+ depth_kernel_size=16,
70
+ depth_num_blocks=12,
71
+ depth_num_heads=8,
72
+ depth_drop_path=0.0,
73
+ thermal_embed_dim=768,
74
+ thermal_kernel_size=16,
75
+ thermal_num_blocks=12,
76
+ thermal_num_heads=12,
77
+ thermal_drop_path=0.0,
78
+ imu_embed_dim=512,
79
+ imu_kernel_size=8,
80
+ imu_num_blocks=6,
81
+ imu_num_heads=8,
82
+ imu_drop_path=0.7,
83
+ ):
84
+ super().__init__()
85
+
86
+ self.modality_preprocessors = self._create_modality_preprocessors(
87
+ video_frames,
88
+ vision_embed_dim,
89
+ kernel_size,
90
+ text_embed_dim,
91
+ audio_embed_dim,
92
+ audio_kernel_size,
93
+ audio_stride,
94
+ audio_num_mel_bins,
95
+ audio_target_len,
96
+ depth_embed_dim,
97
+ depth_kernel_size,
98
+ thermal_embed_dim,
99
+ thermal_kernel_size,
100
+ imu_embed_dim,
101
+ )
102
+
103
+ self.modality_trunks = self._create_modality_trunks(
104
+ vision_embed_dim,
105
+ vision_num_blocks,
106
+ vision_num_heads,
107
+ text_embed_dim,
108
+ text_num_blocks,
109
+ text_num_heads,
110
+ audio_embed_dim,
111
+ audio_num_blocks,
112
+ audio_num_heads,
113
+ audio_drop_path,
114
+ depth_embed_dim,
115
+ depth_num_blocks,
116
+ depth_num_heads,
117
+ depth_drop_path,
118
+ thermal_embed_dim,
119
+ thermal_num_blocks,
120
+ thermal_num_heads,
121
+ thermal_drop_path,
122
+ imu_embed_dim,
123
+ imu_num_blocks,
124
+ imu_num_heads,
125
+ imu_drop_path,
126
+ )
127
+
128
+ self.modality_heads = self._create_modality_heads(
129
+ out_embed_dim,
130
+ vision_embed_dim,
131
+ text_embed_dim,
132
+ audio_embed_dim,
133
+ depth_embed_dim,
134
+ thermal_embed_dim,
135
+ imu_embed_dim,
136
+ )
137
+
138
+ self.modality_postprocessors = self._create_modality_postprocessors(
139
+ out_embed_dim
140
+ )
141
+
142
+ def _create_modality_preprocessors(
143
+ self,
144
+ video_frames=2,
145
+ vision_embed_dim=1024,
146
+ kernel_size=(2, 14, 14),
147
+ text_embed_dim=768,
148
+ audio_embed_dim=768,
149
+ audio_kernel_size=16,
150
+ audio_stride=10,
151
+ audio_num_mel_bins=128,
152
+ audio_target_len=204,
153
+ depth_embed_dim=768,
154
+ depth_kernel_size=16,
155
+ thermal_embed_dim=768,
156
+ thermal_kernel_size=16,
157
+ imu_embed_dim=512,
158
+ ):
159
+ rgbt_stem = PatchEmbedGeneric(
160
+ proj_stem=[
161
+ PadIm2Video(pad_type="repeat", ntimes=2),
162
+ nn.Conv3d(
163
+ in_channels=3,
164
+ kernel_size=kernel_size,
165
+ out_channels=vision_embed_dim,
166
+ stride=kernel_size,
167
+ bias=False,
168
+ ),
169
+ ]
170
+ )
171
+ rgbt_preprocessor = RGBDTPreprocessor(
172
+ img_size=[3, video_frames, 224, 224],
173
+ num_cls_tokens=1,
174
+ pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
175
+ rgbt_stem=rgbt_stem,
176
+ depth_stem=None,
177
+ )
178
+
179
+ text_preprocessor = TextPreprocessor(
180
+ context_length=77,
181
+ vocab_size=49408,
182
+ embed_dim=text_embed_dim,
183
+ causal_masking=True,
184
+ )
185
+
186
+ audio_stem = PatchEmbedGeneric(
187
+ proj_stem=[
188
+ nn.Conv2d(
189
+ in_channels=1,
190
+ kernel_size=audio_kernel_size,
191
+ stride=audio_stride,
192
+ out_channels=audio_embed_dim,
193
+ bias=False,
194
+ ),
195
+ ],
196
+ norm_layer=nn.LayerNorm(normalized_shape=audio_embed_dim),
197
+ )
198
+ audio_preprocessor = AudioPreprocessor(
199
+ img_size=[1, audio_num_mel_bins, audio_target_len],
200
+ num_cls_tokens=1,
201
+ pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
202
+ audio_stem=audio_stem,
203
+ )
204
+
205
+ depth_stem = PatchEmbedGeneric(
206
+ [
207
+ nn.Conv2d(
208
+ kernel_size=depth_kernel_size,
209
+ in_channels=1,
210
+ out_channels=depth_embed_dim,
211
+ stride=depth_kernel_size,
212
+ bias=False,
213
+ ),
214
+ ],
215
+ norm_layer=nn.LayerNorm(normalized_shape=depth_embed_dim),
216
+ )
217
+
218
+ depth_preprocessor = RGBDTPreprocessor(
219
+ img_size=[1, 224, 224],
220
+ num_cls_tokens=1,
221
+ pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
222
+ rgbt_stem=None,
223
+ depth_stem=depth_stem,
224
+ )
225
+
226
+ thermal_stem = PatchEmbedGeneric(
227
+ [
228
+ nn.Conv2d(
229
+ kernel_size=thermal_kernel_size,
230
+ in_channels=1,
231
+ out_channels=thermal_embed_dim,
232
+ stride=thermal_kernel_size,
233
+ bias=False,
234
+ ),
235
+ ],
236
+ norm_layer=nn.LayerNorm(normalized_shape=thermal_embed_dim),
237
+ )
238
+ thermal_preprocessor = ThermalPreprocessor(
239
+ img_size=[1, 224, 224],
240
+ num_cls_tokens=1,
241
+ pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
242
+ thermal_stem=thermal_stem,
243
+ )
244
+
245
+ imu_stem = PatchEmbedGeneric(
246
+ [
247
+ nn.Linear(
248
+ in_features=48,
249
+ out_features=imu_embed_dim,
250
+ bias=False,
251
+ ),
252
+ ],
253
+ norm_layer=nn.LayerNorm(normalized_shape=imu_embed_dim),
254
+ )
255
+
256
+ imu_preprocessor = IMUPreprocessor(
257
+ img_size=[6, 2000],
258
+ num_cls_tokens=1,
259
+ kernel_size=8,
260
+ embed_dim=imu_embed_dim,
261
+ pos_embed_fn=partial(SpatioTemporalPosEmbeddingHelper, learnable=True),
262
+ imu_stem=imu_stem,
263
+ )
264
+
265
+ modality_preprocessors = {
266
+ ModalityType.VISION: rgbt_preprocessor,
267
+ ModalityType.TEXT: text_preprocessor,
268
+ ModalityType.AUDIO: audio_preprocessor,
269
+ ModalityType.DEPTH: depth_preprocessor,
270
+ ModalityType.THERMAL: thermal_preprocessor,
271
+ ModalityType.IMU: imu_preprocessor,
272
+ }
273
+
274
+ return nn.ModuleDict(modality_preprocessors)
275
+
276
+ def _create_modality_trunks(
277
+ self,
278
+ vision_embed_dim=1024,
279
+ vision_num_blocks=24,
280
+ vision_num_heads=16,
281
+ text_embed_dim=768,
282
+ text_num_blocks=12,
283
+ text_num_heads=12,
284
+ audio_embed_dim=768,
285
+ audio_num_blocks=12,
286
+ audio_num_heads=12,
287
+ audio_drop_path=0.0,
288
+ depth_embed_dim=768,
289
+ depth_num_blocks=12,
290
+ depth_num_heads=12,
291
+ depth_drop_path=0.0,
292
+ thermal_embed_dim=768,
293
+ thermal_num_blocks=12,
294
+ thermal_num_heads=12,
295
+ thermal_drop_path=0.0,
296
+ imu_embed_dim=512,
297
+ imu_num_blocks=6,
298
+ imu_num_heads=8,
299
+ imu_drop_path=0.7,
300
+ ):
301
+ def instantiate_trunk(
302
+ embed_dim, num_blocks, num_heads, pre_transformer_ln, add_bias_kv, drop_path
303
+ ):
304
+ return SimpleTransformer(
305
+ embed_dim=embed_dim,
306
+ num_blocks=num_blocks,
307
+ ffn_dropout_rate=0.0,
308
+ drop_path_rate=drop_path,
309
+ attn_target=partial(
310
+ MultiheadAttention,
311
+ embed_dim=embed_dim,
312
+ num_heads=num_heads,
313
+ bias=True,
314
+ add_bias_kv=add_bias_kv,
315
+ ),
316
+ pre_transformer_layer=nn.Sequential(
317
+ nn.LayerNorm(embed_dim, eps=1e-6)
318
+ if pre_transformer_ln
319
+ else nn.Identity(),
320
+ EinOpsRearrange("b l d -> l b d"),
321
+ ),
322
+ post_transformer_layer=EinOpsRearrange("l b d -> b l d"),
323
+ )
324
+
325
+ modality_trunks = {}
326
+ modality_trunks[ModalityType.VISION] = instantiate_trunk(
327
+ vision_embed_dim,
328
+ vision_num_blocks,
329
+ vision_num_heads,
330
+ pre_transformer_ln=True,
331
+ add_bias_kv=False,
332
+ drop_path=0.0,
333
+ )
334
+ modality_trunks[ModalityType.TEXT] = instantiate_trunk(
335
+ text_embed_dim,
336
+ text_num_blocks,
337
+ text_num_heads,
338
+ pre_transformer_ln=False,
339
+ add_bias_kv=False,
340
+ drop_path=0.0,
341
+ )
342
+ modality_trunks[ModalityType.AUDIO] = instantiate_trunk(
343
+ audio_embed_dim,
344
+ audio_num_blocks,
345
+ audio_num_heads,
346
+ pre_transformer_ln=False,
347
+ add_bias_kv=True,
348
+ drop_path=audio_drop_path,
349
+ )
350
+ modality_trunks[ModalityType.DEPTH] = instantiate_trunk(
351
+ depth_embed_dim,
352
+ depth_num_blocks,
353
+ depth_num_heads,
354
+ pre_transformer_ln=False,
355
+ add_bias_kv=True,
356
+ drop_path=depth_drop_path,
357
+ )
358
+ modality_trunks[ModalityType.THERMAL] = instantiate_trunk(
359
+ thermal_embed_dim,
360
+ thermal_num_blocks,
361
+ thermal_num_heads,
362
+ pre_transformer_ln=False,
363
+ add_bias_kv=True,
364
+ drop_path=thermal_drop_path,
365
+ )
366
+ modality_trunks[ModalityType.IMU] = instantiate_trunk(
367
+ imu_embed_dim,
368
+ imu_num_blocks,
369
+ imu_num_heads,
370
+ pre_transformer_ln=False,
371
+ add_bias_kv=True,
372
+ drop_path=imu_drop_path,
373
+ )
374
+
375
+ return nn.ModuleDict(modality_trunks)
376
+
377
+ def _create_modality_heads(
378
+ self,
379
+ out_embed_dim,
380
+ vision_embed_dim,
381
+ text_embed_dim,
382
+ audio_embed_dim,
383
+ depth_embed_dim,
384
+ thermal_embed_dim,
385
+ imu_embed_dim,
386
+ ):
387
+ modality_heads = {}
388
+
389
+ modality_heads[ModalityType.VISION] = nn.Sequential(
390
+ nn.LayerNorm(normalized_shape=vision_embed_dim, eps=1e-6),
391
+ SelectElement(index=0),
392
+ nn.Linear(vision_embed_dim, out_embed_dim, bias=False),
393
+ )
394
+
395
+ modality_heads[ModalityType.TEXT] = SelectEOSAndProject(
396
+ proj=nn.Sequential(
397
+ nn.LayerNorm(normalized_shape=text_embed_dim, eps=1e-6),
398
+ nn.Linear(text_embed_dim, out_embed_dim, bias=False),
399
+ )
400
+ )
401
+
402
+ modality_heads[ModalityType.AUDIO] = nn.Sequential(
403
+ nn.LayerNorm(normalized_shape=audio_embed_dim, eps=1e-6),
404
+ SelectElement(index=0),
405
+ nn.Linear(audio_embed_dim, out_embed_dim, bias=False),
406
+ )
407
+
408
+ modality_heads[ModalityType.DEPTH] = nn.Sequential(
409
+ nn.LayerNorm(normalized_shape=depth_embed_dim, eps=1e-6),
410
+ SelectElement(index=0),
411
+ nn.Linear(depth_embed_dim, out_embed_dim, bias=False),
412
+ )
413
+
414
+ modality_heads[ModalityType.THERMAL] = nn.Sequential(
415
+ nn.LayerNorm(normalized_shape=thermal_embed_dim, eps=1e-6),
416
+ SelectElement(index=0),
417
+ nn.Linear(thermal_embed_dim, out_embed_dim, bias=False),
418
+ )
419
+
420
+ modality_heads[ModalityType.IMU] = nn.Sequential(
421
+ nn.LayerNorm(normalized_shape=imu_embed_dim, eps=1e-6),
422
+ SelectElement(index=0),
423
+ nn.Dropout(p=0.5),
424
+ nn.Linear(imu_embed_dim, out_embed_dim, bias=False),
425
+ )
426
+
427
+ return nn.ModuleDict(modality_heads)
428
+
429
+ def _create_modality_postprocessors(self, out_embed_dim):
430
+ modality_postprocessors = {}
431
+
432
+ modality_postprocessors[ModalityType.VISION] = Normalize(dim=-1)
433
+ modality_postprocessors[ModalityType.TEXT] = nn.Sequential(
434
+ Normalize(dim=-1), LearnableLogitScaling(learnable=True)
435
+ )
436
+ modality_postprocessors[ModalityType.AUDIO] = nn.Sequential(
437
+ Normalize(dim=-1),
438
+ LearnableLogitScaling(logit_scale_init=20.0, learnable=False),
439
+ )
440
+ modality_postprocessors[ModalityType.DEPTH] = nn.Sequential(
441
+ Normalize(dim=-1),
442
+ LearnableLogitScaling(logit_scale_init=5.0, learnable=False),
443
+ )
444
+ modality_postprocessors[ModalityType.THERMAL] = nn.Sequential(
445
+ Normalize(dim=-1),
446
+ LearnableLogitScaling(logit_scale_init=10.0, learnable=False),
447
+ )
448
+ modality_postprocessors[ModalityType.IMU] = nn.Sequential(
449
+ Normalize(dim=-1),
450
+ LearnableLogitScaling(logit_scale_init=5.0, learnable=False),
451
+ )
452
+
453
+ return nn.ModuleDict(modality_postprocessors)
454
+
455
+ def forward(self, inputs):
456
+ outputs = {}
457
+ for modality_key, modality_value in inputs.items():
458
+ reduce_list = (
459
+ modality_value.ndim >= 5
460
+ ) # Audio and Video inputs consist of multiple clips
461
+ if reduce_list:
462
+ B, S = modality_value.shape[:2]
463
+ modality_value = modality_value.reshape(
464
+ B * S, *modality_value.shape[2:]
465
+ )
466
+
467
+ if modality_value is not None:
468
+ modality_value = self.modality_preprocessors[modality_key](
469
+ **{modality_key: modality_value}
470
+ )
471
+ trunk_inputs = modality_value["trunk"]
472
+ head_inputs = modality_value["head"]
473
+ modality_value = self.modality_trunks[modality_key](**trunk_inputs)
474
+ modality_value = self.modality_heads[modality_key](
475
+ modality_value, **head_inputs
476
+ )
477
+ modality_value = self.modality_postprocessors[modality_key](
478
+ modality_value
479
+ )
480
+
481
+ if reduce_list:
482
+ modality_value = modality_value.reshape(B, S, -1)
483
+ modality_value = modality_value.mean(dim=1)
484
+
485
+ outputs[modality_key] = modality_value
486
+
487
+ return outputs
488
+
489
+
490
+ def imagebind_huge(pretrained=False):
491
+ model = ImageBindModel(
492
+ vision_embed_dim=1280,
493
+ vision_num_blocks=32,
494
+ vision_num_heads=16,
495
+ text_embed_dim=1024,
496
+ text_num_blocks=24,
497
+ text_num_heads=16,
498
+ out_embed_dim=1024,
499
+ audio_drop_path=0.1,
500
+ imu_drop_path=0.7,
501
+ )
502
+
503
+ if pretrained:
504
+ if not os.path.exists(".checkpoints/imagebind_huge.pth"):
505
+ print(
506
+ "Downloading imagebind weights to .checkpoints/imagebind_huge.pth ..."
507
+ )
508
+ os.makedirs(".checkpoints", exist_ok=True)
509
+ torch.hub.download_url_to_file(
510
+ "https://dl.fbaipublicfiles.com/imagebind/imagebind_huge.pth",
511
+ ".checkpoints/imagebind_huge.pth",
512
+ progress=True,
513
+ )
514
+
515
+ model.load_state_dict(torch.load(".checkpoints/imagebind_huge.pth"))
516
+
517
+ return model
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/multimodal_preprocessors.py ADDED
@@ -0,0 +1,687 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ # Portions Copyright (c) Meta Platforms, Inc. and affiliates.
3
+ # All rights reserved.
4
+
5
+ # This source code is licensed under the license found in the
6
+ # LICENSE file in the root directory of this source tree.
7
+
8
+ import gzip
9
+ import html
10
+ import io
11
+ import math
12
+ from functools import lru_cache
13
+ from typing import Callable, List, Optional
14
+
15
+ import ftfy
16
+
17
+ import numpy as np
18
+ import regex as re
19
+ import torch
20
+ import torch.nn as nn
21
+ from iopath.common.file_io import g_pathmgr
22
+ from timm.models.layers import trunc_normal_
23
+
24
+ from models.helpers import cast_if_src_dtype, VerboseNNModule
25
+
26
+
27
+ def get_sinusoid_encoding_table(n_position, d_hid):
28
+ """Sinusoid position encoding table"""
29
+
30
+ # TODO: make it with torch instead of numpy
31
+ def get_position_angle_vec(position):
32
+ return [
33
+ position / np.power(10000, 2 * (hid_j // 2) / d_hid)
34
+ for hid_j in range(d_hid)
35
+ ]
36
+
37
+ sinusoid_table = np.array(
38
+ [get_position_angle_vec(pos_i) for pos_i in range(n_position)]
39
+ )
40
+ sinusoid_table[:, 0::2] = np.sin(sinusoid_table[:, 0::2]) # dim 2i
41
+ sinusoid_table[:, 1::2] = np.cos(sinusoid_table[:, 1::2]) # dim 2i+1
42
+
43
+ return torch.FloatTensor(sinusoid_table).unsqueeze(0)
44
+
45
+
46
+ def interpolate_pos_encoding_2d(target_spatial_size, pos_embed):
47
+ N = pos_embed.shape[1]
48
+ if N == target_spatial_size:
49
+ return pos_embed
50
+ dim = pos_embed.shape[-1]
51
+ # nn.functional.interpolate doesn't work with bfloat16 so we cast to float32
52
+ pos_embed, updated = cast_if_src_dtype(pos_embed, torch.bfloat16, torch.float32)
53
+ pos_embed = nn.functional.interpolate(
54
+ pos_embed.reshape(1, int(math.sqrt(N)), int(math.sqrt(N)), dim).permute(
55
+ 0, 3, 1, 2
56
+ ),
57
+ scale_factor=math.sqrt(target_spatial_size / N),
58
+ mode="bicubic",
59
+ )
60
+ if updated:
61
+ pos_embed, _ = cast_if_src_dtype(pos_embed, torch.float32, torch.bfloat16)
62
+ pos_embed = pos_embed.permute(0, 2, 3, 1).view(1, -1, dim)
63
+ return pos_embed
64
+
65
+
66
+ def interpolate_pos_encoding(
67
+ npatch_per_img,
68
+ pos_embed,
69
+ patches_layout,
70
+ input_shape=None,
71
+ first_patch_idx=1,
72
+ ):
73
+ assert first_patch_idx == 0 or first_patch_idx == 1, "there is 1 CLS token or none"
74
+ N = pos_embed.shape[1] - first_patch_idx # since it's 1 if cls_token exists
75
+ if npatch_per_img == N:
76
+ return pos_embed
77
+
78
+ assert (
79
+ patches_layout[-1] == patches_layout[-2]
80
+ ), "Interpolation of pos embed not supported for non-square layouts"
81
+
82
+ class_emb = pos_embed[:, :first_patch_idx]
83
+ pos_embed = pos_embed[:, first_patch_idx:]
84
+
85
+ if input_shape is None or patches_layout[0] == 1:
86
+ # simple 2D pos embedding, no temporal component
87
+ pos_embed = interpolate_pos_encoding_2d(npatch_per_img, pos_embed)
88
+ elif patches_layout[0] > 1:
89
+ # pos embed has a temporal component
90
+ assert len(input_shape) == 4, "temporal interpolation not supported"
91
+ # we only support 2D interpolation in this case
92
+ num_frames = patches_layout[0]
93
+ num_spatial_tokens = patches_layout[1] * patches_layout[2]
94
+ pos_embed = pos_embed.view(1, num_frames, num_spatial_tokens, -1)
95
+ # interpolate embedding for zeroth frame
96
+ pos_embed = interpolate_pos_encoding_2d(
97
+ npatch_per_img, pos_embed[0, 0, ...].unsqueeze(0)
98
+ )
99
+ else:
100
+ raise ValueError("This type of interpolation isn't implemented")
101
+
102
+ return torch.cat((class_emb, pos_embed), dim=1)
103
+
104
+
105
+ def _get_pos_embedding(
106
+ npatch_per_img,
107
+ pos_embed,
108
+ patches_layout,
109
+ input_shape,
110
+ first_patch_idx=1,
111
+ ):
112
+ pos_embed = interpolate_pos_encoding(
113
+ npatch_per_img,
114
+ pos_embed,
115
+ patches_layout,
116
+ input_shape=input_shape,
117
+ first_patch_idx=first_patch_idx,
118
+ )
119
+ return pos_embed
120
+
121
+
122
+ class PatchEmbedGeneric(nn.Module):
123
+ """
124
+ PatchEmbed from Hydra
125
+ """
126
+
127
+ def __init__(self, proj_stem, norm_layer: Optional[nn.Module] = None):
128
+ super().__init__()
129
+
130
+ if len(proj_stem) > 1:
131
+ self.proj = nn.Sequential(*proj_stem)
132
+ else:
133
+ # Special case to be able to load pre-trained models that were
134
+ # trained with a standard stem
135
+ self.proj = proj_stem[0]
136
+ self.norm_layer = norm_layer
137
+
138
+ def get_patch_layout(self, img_size):
139
+ with torch.no_grad():
140
+ dummy_img = torch.zeros(
141
+ [
142
+ 1,
143
+ ]
144
+ + img_size
145
+ )
146
+ dummy_out = self.proj(dummy_img)
147
+ embed_dim = dummy_out.shape[1]
148
+ patches_layout = tuple(dummy_out.shape[2:])
149
+ num_patches = np.prod(patches_layout)
150
+ return patches_layout, num_patches, embed_dim
151
+
152
+ def forward(self, x):
153
+ x = self.proj(x)
154
+ # B C (T) H W -> B (T)HW C
155
+ x = x.flatten(2).transpose(1, 2)
156
+ if self.norm_layer is not None:
157
+ x = self.norm_layer(x)
158
+ return x
159
+
160
+
161
+ class SpatioTemporalPosEmbeddingHelper(VerboseNNModule):
162
+ def __init__(
163
+ self,
164
+ patches_layout: List,
165
+ num_patches: int,
166
+ num_cls_tokens: int,
167
+ embed_dim: int,
168
+ learnable: bool,
169
+ ) -> None:
170
+ super().__init__()
171
+ self.num_cls_tokens = num_cls_tokens
172
+ self.patches_layout = patches_layout
173
+ self.num_patches = num_patches
174
+ self.num_tokens = num_cls_tokens + num_patches
175
+ self.learnable = learnable
176
+ if self.learnable:
177
+ self.pos_embed = nn.Parameter(torch.zeros(1, self.num_tokens, embed_dim))
178
+ trunc_normal_(self.pos_embed, std=0.02)
179
+ else:
180
+ self.register_buffer(
181
+ "pos_embed", get_sinusoid_encoding_table(self.num_tokens, embed_dim)
182
+ )
183
+
184
+ def get_pos_embedding(self, vision_input, all_vision_tokens):
185
+ input_shape = vision_input.shape
186
+ pos_embed = _get_pos_embedding(
187
+ all_vision_tokens.size(1) - self.num_cls_tokens,
188
+ pos_embed=self.pos_embed,
189
+ patches_layout=self.patches_layout,
190
+ input_shape=input_shape,
191
+ first_patch_idx=self.num_cls_tokens,
192
+ )
193
+ return pos_embed
194
+
195
+
196
+ class RGBDTPreprocessor(VerboseNNModule):
197
+ def __init__(
198
+ self,
199
+ rgbt_stem: PatchEmbedGeneric,
200
+ depth_stem: PatchEmbedGeneric,
201
+ img_size: List = (3, 224, 224),
202
+ num_cls_tokens: int = 1,
203
+ pos_embed_fn: Callable = None,
204
+ use_type_embed: bool = False,
205
+ init_param_style: str = "openclip",
206
+ ) -> None:
207
+ super().__init__()
208
+ stem = rgbt_stem if rgbt_stem is not None else depth_stem
209
+ (
210
+ self.patches_layout,
211
+ self.num_patches,
212
+ self.embed_dim,
213
+ ) = stem.get_patch_layout(img_size)
214
+ self.rgbt_stem = rgbt_stem
215
+ self.depth_stem = depth_stem
216
+ self.use_pos_embed = pos_embed_fn is not None
217
+ self.use_type_embed = use_type_embed
218
+ self.num_cls_tokens = num_cls_tokens
219
+
220
+ if self.use_pos_embed:
221
+ self.pos_embedding_helper = pos_embed_fn(
222
+ patches_layout=self.patches_layout,
223
+ num_cls_tokens=num_cls_tokens,
224
+ num_patches=self.num_patches,
225
+ embed_dim=self.embed_dim,
226
+ )
227
+ if self.num_cls_tokens > 0:
228
+ self.cls_token = nn.Parameter(
229
+ torch.zeros(1, self.num_cls_tokens, self.embed_dim)
230
+ )
231
+ if self.use_type_embed:
232
+ self.type_embed = nn.Parameter(torch.zeros(1, 1, self.embed_dim))
233
+
234
+ self.init_parameters(init_param_style)
235
+
236
+ @torch.no_grad()
237
+ def init_parameters(self, init_param_style):
238
+ if init_param_style == "openclip":
239
+ # OpenCLIP style initialization
240
+ scale = self.embed_dim**-0.5
241
+ if self.use_pos_embed:
242
+ nn.init.normal_(self.pos_embedding_helper.pos_embed)
243
+ self.pos_embedding_helper.pos_embed *= scale
244
+
245
+ if self.num_cls_tokens > 0:
246
+ nn.init.normal_(self.cls_token)
247
+ self.cls_token *= scale
248
+ elif init_param_style == "vit":
249
+ self.cls_token.data.fill_(0)
250
+ else:
251
+ raise ValueError(f"Unknown init {init_param_style}")
252
+
253
+ if self.use_type_embed:
254
+ nn.init.normal_(self.type_embed)
255
+
256
+ def tokenize_input_and_cls_pos(self, input, stem, mask):
257
+ # tokens is of shape B x L x D
258
+ tokens = stem(input)
259
+ assert tokens.ndim == 3
260
+ assert tokens.shape[2] == self.embed_dim
261
+ B = tokens.shape[0]
262
+ if self.num_cls_tokens > 0:
263
+ class_tokens = self.cls_token.expand(
264
+ B, -1, -1
265
+ ) # stole class_tokens impl from Phil Wang, thanks
266
+ tokens = torch.cat((class_tokens, tokens), dim=1)
267
+ if self.use_pos_embed:
268
+ pos_embed = self.pos_embedding_helper.get_pos_embedding(input, tokens)
269
+ tokens = tokens + pos_embed
270
+ if self.use_type_embed:
271
+ tokens = tokens + self.type_embed.expand(B, -1, -1)
272
+ return tokens
273
+
274
+ def forward(self, vision=None, depth=None, patch_mask=None):
275
+ if patch_mask is not None:
276
+ raise NotImplementedError()
277
+
278
+ if vision is not None:
279
+ vision_tokens = self.tokenize_input_and_cls_pos(
280
+ vision, self.rgbt_stem, patch_mask
281
+ )
282
+
283
+ if depth is not None:
284
+ depth_tokens = self.tokenize_input_and_cls_pos(
285
+ depth, self.depth_stem, patch_mask
286
+ )
287
+
288
+ # aggregate tokens
289
+ if vision is not None and depth is not None:
290
+ final_tokens = vision_tokens + depth_tokens
291
+ else:
292
+ final_tokens = vision_tokens if vision is not None else depth_tokens
293
+ return_dict = {
294
+ "trunk": {
295
+ "tokens": final_tokens,
296
+ },
297
+ "head": {},
298
+ }
299
+ return return_dict
300
+
301
+
302
+ class AudioPreprocessor(RGBDTPreprocessor):
303
+ def __init__(self, audio_stem: PatchEmbedGeneric, **kwargs) -> None:
304
+ super().__init__(rgbt_stem=audio_stem, depth_stem=None, **kwargs)
305
+
306
+ def forward(self, audio=None):
307
+ return super().forward(vision=audio)
308
+
309
+
310
+ class ThermalPreprocessor(RGBDTPreprocessor):
311
+ def __init__(self, thermal_stem: PatchEmbedGeneric, **kwargs) -> None:
312
+ super().__init__(rgbt_stem=thermal_stem, depth_stem=None, **kwargs)
313
+
314
+ def forward(self, thermal=None):
315
+ return super().forward(vision=thermal)
316
+
317
+
318
+ def build_causal_attention_mask(context_length):
319
+ # lazily create causal attention mask, with full attention between the vision tokens
320
+ # pytorch uses additive attention mask; fill with -inf
321
+ mask = torch.empty(context_length, context_length, requires_grad=False)
322
+ mask.fill_(float("-inf"))
323
+ mask.triu_(1) # zero out the lower diagonal
324
+ return mask
325
+
326
+
327
+ class TextPreprocessor(VerboseNNModule):
328
+ def __init__(
329
+ self,
330
+ vocab_size: int,
331
+ context_length: int,
332
+ embed_dim: int,
333
+ causal_masking: bool,
334
+ supply_seq_len_to_head: bool = True,
335
+ num_cls_tokens: int = 0,
336
+ init_param_style: str = "openclip",
337
+ ) -> None:
338
+ super().__init__()
339
+ self.vocab_size = vocab_size
340
+ self.context_length = context_length
341
+ self.token_embedding = nn.Embedding(vocab_size, embed_dim)
342
+ self.pos_embed = nn.Parameter(
343
+ torch.empty(1, self.context_length + num_cls_tokens, embed_dim)
344
+ )
345
+ self.causal_masking = causal_masking
346
+ if self.causal_masking:
347
+ mask = build_causal_attention_mask(self.context_length)
348
+ # register the mask as a buffer so it can be moved to the right device
349
+ self.register_buffer("mask", mask)
350
+
351
+ self.supply_seq_len_to_head = supply_seq_len_to_head
352
+ self.num_cls_tokens = num_cls_tokens
353
+ self.embed_dim = embed_dim
354
+ if num_cls_tokens > 0:
355
+ assert self.causal_masking is False, "Masking + CLS token isn't implemented"
356
+ self.cls_token = nn.Parameter(
357
+ torch.zeros(1, self.num_cls_tokens, embed_dim)
358
+ )
359
+
360
+ self.init_parameters(init_param_style)
361
+
362
+ @torch.no_grad()
363
+ def init_parameters(self, init_param_style="openclip"):
364
+ # OpenCLIP style initialization
365
+ nn.init.normal_(self.token_embedding.weight, std=0.02)
366
+ nn.init.normal_(self.pos_embed, std=0.01)
367
+
368
+ if init_param_style == "openclip":
369
+ # OpenCLIP style initialization
370
+ scale = self.embed_dim**-0.5
371
+ if self.num_cls_tokens > 0:
372
+ nn.init.normal_(self.cls_token)
373
+ self.cls_token *= scale
374
+ elif init_param_style == "vit":
375
+ self.cls_token.data.fill_(0)
376
+ else:
377
+ raise ValueError(f"Unknown init {init_param_style}")
378
+
379
+ def forward(self, text):
380
+ # text tokens are of shape B x L x D
381
+ text_tokens = self.token_embedding(text)
382
+ # concat CLS tokens if any
383
+ if self.num_cls_tokens > 0:
384
+ B = text_tokens.shape[0]
385
+ class_tokens = self.cls_token.expand(
386
+ B, -1, -1
387
+ ) # stole class_tokens impl from Phil Wang, thanks
388
+ text_tokens = torch.cat((class_tokens, text_tokens), dim=1)
389
+ text_tokens = text_tokens + self.pos_embed
390
+ return_dict = {
391
+ "trunk": {
392
+ "tokens": text_tokens,
393
+ },
394
+ "head": {},
395
+ }
396
+ # Compute sequence length after adding CLS tokens
397
+ if self.supply_seq_len_to_head:
398
+ text_lengths = text.argmax(dim=-1)
399
+ return_dict["head"] = {
400
+ "seq_len": text_lengths,
401
+ }
402
+ if self.causal_masking:
403
+ return_dict["trunk"].update({"attn_mask": self.mask})
404
+ return return_dict
405
+
406
+
407
+ class Im2Video(nn.Module):
408
+ """Convert an image into a trivial video."""
409
+
410
+ def __init__(self, time_dim=2):
411
+ super().__init__()
412
+ self.time_dim = time_dim
413
+
414
+ def forward(self, x):
415
+ if x.ndim == 4:
416
+ # B, C, H, W -> B, C, T, H, W
417
+ return x.unsqueeze(self.time_dim)
418
+ elif x.ndim == 5:
419
+ return x
420
+ else:
421
+ raise ValueError(f"Dimension incorrect {x.shape}")
422
+
423
+
424
+ class PadIm2Video(Im2Video):
425
+ def __init__(self, ntimes, pad_type, time_dim=2):
426
+ super().__init__(time_dim=time_dim)
427
+ assert ntimes > 0
428
+ assert pad_type in ["zero", "repeat"]
429
+ self.ntimes = ntimes
430
+ self.pad_type = pad_type
431
+
432
+ def forward(self, x):
433
+ x = super().forward(x)
434
+ if x.shape[self.time_dim] == 1:
435
+ if self.pad_type == "repeat":
436
+ new_shape = [1] * len(x.shape)
437
+ new_shape[self.time_dim] = self.ntimes
438
+ x = x.repeat(new_shape)
439
+ elif self.pad_type == "zero":
440
+ padarg = [0, 0] * len(x.shape)
441
+ padarg[2 * self.time_dim + 1] = self.ntimes - x.shape[self.time_dim]
442
+ x = nn.functional.pad(x, padarg)
443
+ return x
444
+
445
+
446
+ # Modified from github.com/openai/CLIP
447
+ @lru_cache()
448
+ def bytes_to_unicode():
449
+ """
450
+ Returns list of utf-8 byte and a corresponding list of unicode strings.
451
+ The reversible bpe codes work on unicode strings.
452
+ This means you need a large # of unicode characters in your vocab if you want to avoid UNKs.
453
+ When you're at something like a 10B token dataset you end up needing around 5K for decent coverage.
454
+ This is a signficant percentage of your normal, say, 32K bpe vocab.
455
+ To avoid that, we want lookup tables between utf-8 bytes and unicode strings.
456
+ And avoids mapping to whitespace/control characters the bpe code barfs on.
457
+ """
458
+ bs = (
459
+ list(range(ord("!"), ord("~") + 1))
460
+ + list(range(ord("¡"), ord("¬") + 1))
461
+ + list(range(ord("®"), ord("ÿ") + 1))
462
+ )
463
+ cs = bs[:]
464
+ n = 0
465
+ for b in range(2**8):
466
+ if b not in bs:
467
+ bs.append(b)
468
+ cs.append(2**8 + n)
469
+ n += 1
470
+ cs = [chr(n) for n in cs]
471
+ return dict(zip(bs, cs))
472
+
473
+
474
+ def get_pairs(word):
475
+ """Return set of symbol pairs in a word.
476
+ Word is represented as tuple of symbols (symbols being variable-length strings).
477
+ """
478
+ pairs = set()
479
+ prev_char = word[0]
480
+ for char in word[1:]:
481
+ pairs.add((prev_char, char))
482
+ prev_char = char
483
+ return pairs
484
+
485
+
486
+ def basic_clean(text):
487
+ text = ftfy.fix_text(text)
488
+ text = html.unescape(html.unescape(text))
489
+ return text.strip()
490
+
491
+
492
+ def whitespace_clean(text):
493
+ text = re.sub(r"\s+", " ", text)
494
+ text = text.strip()
495
+ return text
496
+
497
+
498
+ class SimpleTokenizer(object):
499
+ def __init__(self, bpe_path: str, context_length=77):
500
+ self.byte_encoder = bytes_to_unicode()
501
+ self.byte_decoder = {v: k for k, v in self.byte_encoder.items()}
502
+
503
+ with g_pathmgr.open(bpe_path, "rb") as fh:
504
+ bpe_bytes = io.BytesIO(fh.read())
505
+ merges = gzip.open(bpe_bytes).read().decode("utf-8").split("\n")
506
+ merges = merges[1 : 49152 - 256 - 2 + 1]
507
+ merges = [tuple(merge.split()) for merge in merges]
508
+ vocab = list(bytes_to_unicode().values())
509
+ vocab = vocab + [v + "</w>" for v in vocab]
510
+ for merge in merges:
511
+ vocab.append("".join(merge))
512
+ vocab.extend(["<|startoftext|>", "<|endoftext|>"])
513
+ self.encoder = dict(zip(vocab, range(len(vocab))))
514
+ self.decoder = {v: k for k, v in self.encoder.items()}
515
+ self.bpe_ranks = dict(zip(merges, range(len(merges))))
516
+ self.cache = {
517
+ "<|startoftext|>": "<|startoftext|>",
518
+ "<|endoftext|>": "<|endoftext|>",
519
+ }
520
+ self.pat = re.compile(
521
+ r"""<\|startoftext\|>|<\|endoftext\|>|'s|'t|'re|'ve|'m|'ll|'d|[\p{L}]+|[\p{N}]|[^\s\p{L}\p{N}]+""",
522
+ re.IGNORECASE,
523
+ )
524
+ self.context_length = context_length
525
+
526
+ def bpe(self, token):
527
+ if token in self.cache:
528
+ return self.cache[token]
529
+ word = tuple(token[:-1]) + (token[-1] + "</w>",)
530
+ pairs = get_pairs(word)
531
+
532
+ if not pairs:
533
+ return token + "</w>"
534
+
535
+ while True:
536
+ bigram = min(pairs, key=lambda pair: self.bpe_ranks.get(pair, float("inf")))
537
+ if bigram not in self.bpe_ranks:
538
+ break
539
+ first, second = bigram
540
+ new_word = []
541
+ i = 0
542
+ while i < len(word):
543
+ try:
544
+ j = word.index(first, i)
545
+ new_word.extend(word[i:j])
546
+ i = j
547
+ except:
548
+ new_word.extend(word[i:])
549
+ break
550
+
551
+ if word[i] == first and i < len(word) - 1 and word[i + 1] == second:
552
+ new_word.append(first + second)
553
+ i += 2
554
+ else:
555
+ new_word.append(word[i])
556
+ i += 1
557
+ new_word = tuple(new_word)
558
+ word = new_word
559
+ if len(word) == 1:
560
+ break
561
+ else:
562
+ pairs = get_pairs(word)
563
+ word = " ".join(word)
564
+ self.cache[token] = word
565
+ return word
566
+
567
+ def encode(self, text):
568
+ bpe_tokens = []
569
+ text = whitespace_clean(basic_clean(text)).lower()
570
+ for token in re.findall(self.pat, text):
571
+ token = "".join(self.byte_encoder[b] for b in token.encode("utf-8"))
572
+ bpe_tokens.extend(
573
+ self.encoder[bpe_token] for bpe_token in self.bpe(token).split(" ")
574
+ )
575
+ return bpe_tokens
576
+
577
+ def decode(self, tokens):
578
+ text = "".join([self.decoder[token] for token in tokens])
579
+ text = (
580
+ bytearray([self.byte_decoder[c] for c in text])
581
+ .decode("utf-8", errors="replace")
582
+ .replace("</w>", " ")
583
+ )
584
+ return text
585
+
586
+ def __call__(self, texts, context_length=None):
587
+ if not context_length:
588
+ context_length = self.context_length
589
+
590
+ if isinstance(texts, str):
591
+ texts = [texts]
592
+
593
+ sot_token = self.encoder["<|startoftext|>"]
594
+ eot_token = self.encoder["<|endoftext|>"]
595
+ all_tokens = [[sot_token] + self.encode(text) + [eot_token] for text in texts]
596
+ result = torch.zeros(len(all_tokens), context_length, dtype=torch.long)
597
+
598
+ for i, tokens in enumerate(all_tokens):
599
+ tokens = tokens[:context_length]
600
+ result[i, : len(tokens)] = torch.tensor(tokens)
601
+
602
+ if len(result) == 1:
603
+ return result[0]
604
+ return result
605
+
606
+
607
+ class IMUPreprocessor(VerboseNNModule):
608
+ def __init__(
609
+ self,
610
+ kernel_size: int,
611
+ imu_stem: PatchEmbedGeneric,
612
+ embed_dim: int,
613
+ img_size: List = (6, 2000),
614
+ num_cls_tokens: int = 1,
615
+ pos_embed_fn: Callable = None,
616
+ init_param_style: str = "openclip",
617
+ ) -> None:
618
+ super().__init__()
619
+ stem = imu_stem
620
+ self.imu_stem = imu_stem
621
+ self.embed_dim = embed_dim
622
+ self.use_pos_embed = pos_embed_fn is not None
623
+ self.num_cls_tokens = num_cls_tokens
624
+ self.kernel_size = kernel_size
625
+ self.pos_embed = nn.Parameter(
626
+ torch.empty(1, (img_size[1] // kernel_size) + num_cls_tokens, embed_dim)
627
+ )
628
+
629
+ if self.num_cls_tokens > 0:
630
+ self.cls_token = nn.Parameter(
631
+ torch.zeros(1, self.num_cls_tokens, self.embed_dim)
632
+ )
633
+
634
+ self.init_parameters(init_param_style)
635
+
636
+ @torch.no_grad()
637
+ def init_parameters(self, init_param_style):
638
+ nn.init.normal_(self.pos_embed, std=0.01)
639
+
640
+ if init_param_style == "openclip":
641
+ # OpenCLIP style initialization
642
+ scale = self.embed_dim**-0.5
643
+
644
+ if self.num_cls_tokens > 0:
645
+ nn.init.normal_(self.cls_token)
646
+ self.cls_token *= scale
647
+ elif init_param_style == "vit":
648
+ self.cls_token.data.fill_(0)
649
+ else:
650
+ raise ValueError(f"Unknown init {init_param_style}")
651
+
652
+ def tokenize_input_and_cls_pos(self, input, stem):
653
+ # tokens is of shape B x L x D
654
+ tokens = stem.norm_layer(stem.proj(input))
655
+ assert tokens.ndim == 3
656
+ assert tokens.shape[2] == self.embed_dim
657
+ B = tokens.shape[0]
658
+ if self.num_cls_tokens > 0:
659
+ class_tokens = self.cls_token.expand(
660
+ B, -1, -1
661
+ ) # stole class_tokens impl from Phil Wang, thanks
662
+ tokens = torch.cat((class_tokens, tokens), dim=1)
663
+ if self.use_pos_embed:
664
+ tokens = tokens + self.pos_embed
665
+ return tokens
666
+
667
+ def forward(self, imu):
668
+ # Patchify
669
+ imu = imu.unfold(
670
+ -1,
671
+ self.kernel_size,
672
+ self.kernel_size,
673
+ ).permute(0, 2, 1, 3)
674
+ imu = imu.reshape(imu.size(0), imu.size(1), -1)
675
+
676
+ imu_tokens = self.tokenize_input_and_cls_pos(
677
+ imu,
678
+ self.imu_stem,
679
+ )
680
+
681
+ return_dict = {
682
+ "trunk": {
683
+ "tokens": imu_tokens,
684
+ },
685
+ "head": {},
686
+ }
687
+ return return_dict
external/Grounded-Segment-Anything/playground/ImageBind_SAM/models/transformer.py ADDED
@@ -0,0 +1,284 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ #!/usr/bin/env python3
2
+ # Portions Copyright (c) Meta Platforms, Inc. and affiliates.
3
+ # All rights reserved.
4
+
5
+ # This source code is licensed under the license found in the
6
+ # LICENSE file in the root directory of this source tree.
7
+
8
+ # Code modified from
9
+ # https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py ;
10
+ # https://github.com/facebookresearch/deit/blob/main/models.py
11
+ # and https://github.com/facebookresearch/vissl/blob/main/vissl/models/trunks/vision_transformer.py
12
+
13
+
14
+ import copy
15
+ import fnmatch
16
+ import logging
17
+ from functools import partial
18
+ from typing import Callable, List
19
+
20
+ import torch
21
+ import torch.nn as nn
22
+ import torch.utils.checkpoint as checkpoint
23
+
24
+ from timm.models.layers import DropPath, trunc_normal_
25
+
26
+
27
+ class Attention(nn.Module):
28
+ def __init__(
29
+ self,
30
+ dim,
31
+ num_heads=8,
32
+ qkv_bias=False,
33
+ qk_scale=None,
34
+ attn_drop=0.0,
35
+ proj_drop=0.0,
36
+ ):
37
+ super().__init__()
38
+ self.num_heads = num_heads
39
+ head_dim = dim // num_heads
40
+ # NOTE scale factor was wrong in my original version,
41
+ # can set manually to be compat with prev weights
42
+ self.scale = qk_scale or head_dim**-0.5
43
+
44
+ self.qkv = nn.Linear(dim, dim * 3, bias=qkv_bias)
45
+ self.attn_drop = nn.Dropout(attn_drop)
46
+ self.proj = nn.Linear(dim, dim)
47
+ self.proj_drop = nn.Dropout(proj_drop)
48
+
49
+ def forward(self, x):
50
+ B, N, C = x.shape
51
+ qkv = (
52
+ self.qkv(x)
53
+ .reshape(B, N, 3, self.num_heads, C // self.num_heads)
54
+ .permute(2, 0, 3, 1, 4)
55
+ )
56
+ q, k, v = (
57
+ qkv[0],
58
+ qkv[1],
59
+ qkv[2],
60
+ ) # make torchscript happy (cannot use tensor as tuple)
61
+
62
+ attn = (q @ k.transpose(-2, -1)) * self.scale
63
+ attn = attn.softmax(dim=-1)
64
+ attn = self.attn_drop(attn)
65
+
66
+ x = (attn @ v).transpose(1, 2).reshape(B, N, C)
67
+ x = self.proj(x)
68
+ x = self.proj_drop(x)
69
+ return x
70
+
71
+
72
+ class Mlp(nn.Module):
73
+ def __init__(
74
+ self,
75
+ in_features,
76
+ hidden_features=None,
77
+ out_features=None,
78
+ act_layer=nn.GELU,
79
+ drop=0.0,
80
+ ):
81
+ super().__init__()
82
+ out_features = out_features or in_features
83
+ hidden_features = hidden_features or in_features
84
+ self.fc1 = nn.Linear(in_features, hidden_features)
85
+ self.act = act_layer()
86
+ self.fc2 = nn.Linear(hidden_features, out_features)
87
+ self.drop = nn.Dropout(drop)
88
+
89
+ def forward(self, x):
90
+ x = self.fc1(x)
91
+ x = self.act(x)
92
+ x = self.drop(x)
93
+ x = self.fc2(x)
94
+ x = self.drop(x)
95
+ return x
96
+
97
+
98
+ class MultiheadAttention(nn.MultiheadAttention):
99
+ def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
100
+ return super().forward(x, x, x, need_weights=False, attn_mask=attn_mask)[0]
101
+
102
+
103
+ class ViTAttention(Attention):
104
+ def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
105
+ assert attn_mask is None
106
+ return super().forward(x)
107
+
108
+
109
+ class BlockWithMasking(nn.Module):
110
+ def __init__(
111
+ self,
112
+ dim: int,
113
+ attn_target: Callable,
114
+ mlp_ratio: int = 4,
115
+ act_layer: Callable = nn.GELU,
116
+ norm_layer: Callable = nn.LayerNorm,
117
+ ffn_dropout_rate: float = 0.0,
118
+ drop_path: float = 0.0,
119
+ layer_scale_type: str = None,
120
+ layer_scale_init_value: float = 1e-4,
121
+ ):
122
+ super().__init__()
123
+
124
+ assert not isinstance(
125
+ attn_target, nn.Module
126
+ ), "attn_target should be a Callable. Otherwise attn_target is shared across blocks!"
127
+ self.attn = attn_target()
128
+ if drop_path > 0.0:
129
+ self.drop_path = DropPath(drop_path)
130
+ else:
131
+ self.drop_path = nn.Identity()
132
+ self.norm_1 = norm_layer(dim)
133
+ mlp_hidden_dim = int(mlp_ratio * dim)
134
+ self.mlp = Mlp(
135
+ in_features=dim,
136
+ hidden_features=mlp_hidden_dim,
137
+ act_layer=act_layer,
138
+ drop=ffn_dropout_rate,
139
+ )
140
+ self.norm_2 = norm_layer(dim)
141
+ self.layer_scale_type = layer_scale_type
142
+ if self.layer_scale_type is not None:
143
+ assert self.layer_scale_type in [
144
+ "per_channel",
145
+ "scalar",
146
+ ], f"Found Layer scale type {self.layer_scale_type}"
147
+ if self.layer_scale_type == "per_channel":
148
+ # one gamma value per channel
149
+ gamma_shape = [1, 1, dim]
150
+ elif self.layer_scale_type == "scalar":
151
+ # single gamma value for all channels
152
+ gamma_shape = [1, 1, 1]
153
+ # two gammas: for each part of the fwd in the encoder
154
+ self.layer_scale_gamma1 = nn.Parameter(
155
+ torch.ones(size=gamma_shape) * layer_scale_init_value,
156
+ requires_grad=True,
157
+ )
158
+ self.layer_scale_gamma2 = nn.Parameter(
159
+ torch.ones(size=gamma_shape) * layer_scale_init_value,
160
+ requires_grad=True,
161
+ )
162
+
163
+ def forward(self, x: torch.Tensor, attn_mask: torch.Tensor):
164
+ if self.layer_scale_type is None:
165
+ x = x + self.drop_path(self.attn(self.norm_1(x), attn_mask))
166
+ x = x + self.drop_path(self.mlp(self.norm_2(x)))
167
+ else:
168
+ x = (
169
+ x
170
+ + self.drop_path(self.attn(self.norm_1(x), attn_mask))
171
+ * self.layer_scale_gamma1
172
+ )
173
+ x = x + self.drop_path(self.mlp(self.norm_2(x))) * self.layer_scale_gamma2
174
+ return x
175
+
176
+
177
+ _LAYER_NORM = partial(nn.LayerNorm, eps=1e-6)
178
+
179
+
180
+ class SimpleTransformer(nn.Module):
181
+ def __init__(
182
+ self,
183
+ attn_target: Callable,
184
+ embed_dim: int,
185
+ num_blocks: int,
186
+ block: Callable = BlockWithMasking,
187
+ pre_transformer_layer: Callable = None,
188
+ post_transformer_layer: Callable = None,
189
+ drop_path_rate: float = 0.0,
190
+ drop_path_type: str = "progressive",
191
+ norm_layer: Callable = _LAYER_NORM,
192
+ mlp_ratio: int = 4,
193
+ ffn_dropout_rate: float = 0.0,
194
+ layer_scale_type: str = None, # from cait; possible values are None, "per_channel", "scalar"
195
+ layer_scale_init_value: float = 1e-4, # from cait; float
196
+ weight_init_style: str = "jax", # possible values jax or pytorch
197
+ ):
198
+ """
199
+ Simple Transformer with the following features
200
+ 1. Supports masked attention
201
+ 2. Supports DropPath
202
+ 3. Supports LayerScale
203
+ 4. Supports Dropout in Attention and FFN
204
+ 5. Makes few assumptions about the input except that it is a Tensor
205
+ """
206
+ super().__init__()
207
+ self.pre_transformer_layer = pre_transformer_layer
208
+ if drop_path_type == "progressive":
209
+ dpr = [x.item() for x in torch.linspace(0, drop_path_rate, num_blocks)]
210
+ elif drop_path_type == "uniform":
211
+ dpr = [drop_path_rate for i in range(num_blocks)]
212
+ else:
213
+ raise ValueError(f"Unknown drop_path_type: {drop_path_type}")
214
+
215
+ self.blocks = nn.Sequential(
216
+ *[
217
+ block(
218
+ dim=embed_dim,
219
+ attn_target=attn_target,
220
+ mlp_ratio=mlp_ratio,
221
+ ffn_dropout_rate=ffn_dropout_rate,
222
+ drop_path=dpr[i],
223
+ norm_layer=norm_layer,
224
+ layer_scale_type=layer_scale_type,
225
+ layer_scale_init_value=layer_scale_init_value,
226
+ )
227
+ for i in range(num_blocks)
228
+ ]
229
+ )
230
+ self.post_transformer_layer = post_transformer_layer
231
+ self.weight_init_style = weight_init_style
232
+ self.apply(self._init_weights)
233
+
234
+ def _init_weights(self, m):
235
+ if isinstance(m, nn.Linear):
236
+ if self.weight_init_style == "jax":
237
+ # Based on MAE and official Jax ViT implementation
238
+ torch.nn.init.xavier_uniform_(m.weight)
239
+ elif self.weight_init_style == "pytorch":
240
+ # PyTorch ViT uses trunc_normal_
241
+ trunc_normal_(m.weight, std=0.02)
242
+
243
+ if m.bias is not None:
244
+ nn.init.constant_(m.bias, 0)
245
+ elif isinstance(m, (nn.LayerNorm)):
246
+ nn.init.constant_(m.bias, 0)
247
+ nn.init.constant_(m.weight, 1.0)
248
+
249
+ def forward(
250
+ self,
251
+ tokens: torch.Tensor,
252
+ attn_mask: torch.Tensor = None,
253
+ use_checkpoint: bool = False,
254
+ checkpoint_every_n: int = 1,
255
+ checkpoint_blk_ids: List[int] = None,
256
+ ):
257
+ """
258
+ Inputs
259
+ - tokens: data of shape N x L x D (or L x N x D depending on the attention implementation)
260
+ - attn: mask of shape L x L
261
+
262
+ Output
263
+ - x: data of shape N x L x D (or L x N x D depending on the attention implementation)
264
+ """
265
+ if self.pre_transformer_layer:
266
+ tokens = self.pre_transformer_layer(tokens)
267
+ if use_checkpoint and checkpoint_blk_ids is None:
268
+ checkpoint_blk_ids = [
269
+ blk_id
270
+ for blk_id in range(len(self.blocks))
271
+ if blk_id % checkpoint_every_n == 0
272
+ ]
273
+ if checkpoint_blk_ids:
274
+ checkpoint_blk_ids = set(checkpoint_blk_ids)
275
+ for blk_id, blk in enumerate(self.blocks):
276
+ if use_checkpoint and blk_id in checkpoint_blk_ids:
277
+ tokens = checkpoint.checkpoint(
278
+ blk, tokens, attn_mask, use_reentrant=False
279
+ )
280
+ else:
281
+ tokens = blk(tokens, attn_mask=attn_mask)
282
+ if self.post_transformer_layer:
283
+ tokens = self.post_transformer_layer(tokens)
284
+ return tokens
external/Grounded-Segment-Anything/playground/PaintByExample/README.md ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## Paint by Example: Exemplar-based Image Editing with Diffusion Models
2
+
3
+ :grapes: [[Official Project Page](https://github.com/Fantasy-Studio/Paint-by-Example)] &nbsp; :apple:[[Official Online Demo](https://huggingface.co/spaces/Fantasy-Studio/Paint-by-Example)]
4
+
5
+ <div align="center">
6
+
7
+ ![](https://github.com/Fantasy-Studio/Paint-by-Example/blob/main/figure/teaser.png?raw=True)
8
+
9
+ </div>
10
+
11
+ ## Abstract
12
+
13
+ > Language-guided image editing has achieved great success recently. In this paper, for the first time, we investigate exemplar-guided image editing for more precise control. We achieve this goal by leveraging self-supervised training to disentangle and re-organize the source image and the exemplar. However, the naive approach will cause obvious fusing artifacts. We carefully analyze it and propose an information bottleneck and strong augmentations to avoid the trivial solution of directly copying and pasting the exemplar image. Meanwhile, to ensure the controllability of the editing process, we design an arbitrary shape mask for the exemplar image and leverage the classifier-free guidance to increase the similarity to the exemplar image. The whole framework involves a single forward of the diffusion model without any iterative optimization. We demonstrate that our method achieves an impressive performance and enables controllable editing on in-the-wild images with high fidelity.
14
+
15
+ ## Table of Contents
16
+ - [Installation](#installation)
17
+ - [Paint-By-Example Demos](#paint-by-example-demos)
18
+ - [Diffuser Demo](#paintbyexample-diffuser-demos)
19
+ - [PaintByExample with SAM](#paintbyexample-with-sam)
20
+
21
+
22
+ ## TODO
23
+ - [x] PaintByExample Diffuser Demo
24
+ - [x] PaintByExample with SAM
25
+ - [ ] PaintByExample with GroundingDINO
26
+ - [ ] PaintByExample with Grounded-SAM
27
+
28
+ ## Installation
29
+ We're using PaintByExample with diffusers, install diffusers as follows:
30
+ ```bash
31
+ pip install diffusers==0.16.1
32
+ ```
33
+ Then install Grounded-SAM follows [Grounded-SAM Installation](https://github.com/IDEA-Research/Grounded-Segment-Anything#installation) for some extension demos.
34
+
35
+ ## Paint-By-Example Demos
36
+ Here we provide the demos for `PaintByExample`
37
+
38
+
39
+ ### PaintByExample Diffuser Demos
40
+ ```python
41
+ cd playground/PaintByExample
42
+ python paint_by_example.py
43
+ ```
44
+ **Notes:** set `cache_dir` to save the pretrained weights to specific folder. The paint result will be save as `paint_by_example_demo.jpg`:
45
+
46
+ <div align="center">
47
+
48
+ | Input Image | Mask | Example Image | Inpaint Result |
49
+ |:----:|:----:|:----:|:----:|
50
+ | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/input_image.png?raw=true) | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/mask.png?raw=true) | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/example_image.jpg?raw=true" width=55%></div> | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/paint_by_example_demo.jpg?raw=true) |
51
+ | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/input_image.png?raw=true) | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/mask.png?raw=true) | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/pomeranian_example.jpg?raw=true" width=55%></div> | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/paint_by_pomeranian_demo.jpg?raw=true) |
52
+
53
+ </div>
54
+
55
+ ### PaintByExample with SAM
56
+
57
+ In this demo, we did inpaint task by:
58
+ 1. Generate mask by SAM with prompt (box or point)
59
+ 2. Inpaint with mask and example image
60
+
61
+ ```python
62
+ cd playground/PaintByExample
63
+ python sam_paint_by_example.py
64
+ ```
65
+ **Notes:** We set a more `num_inference_steps` (like 200 to 500) to get higher quality image. And we've found that the mask region can influence a lot on the final result (like a panda can not be well inpainted with a region like dog). It needed to have more test on it.
66
+
67
+ | Input Image | SAM Output | Example Image | Inpaint Result |
68
+ |:----:|:----:|:----:|:----:|
69
+ | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/input_image.png?raw=true) | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/demo_with_point_prompt.png?raw=true) | <div style="text-align: center"> <img src="https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/labrador_example.jpg?raw=true" width=55%></div> | ![](https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/sam_paint_by_example_demo.jpg?raw=true) |
70
+
external/Grounded-Segment-Anything/playground/PaintByExample/paint_by_example.py ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # !pip install diffusers transformers
2
+
3
+ import PIL
4
+ import requests
5
+ import torch
6
+ from io import BytesIO
7
+ from diffusers import DiffusionPipeline
8
+
9
+
10
+ """
11
+ Step 1: Download demo images
12
+ """
13
+ def download_image(url):
14
+ response = requests.get(url)
15
+ return PIL.Image.open(BytesIO(response.content)).convert("RGB")
16
+
17
+
18
+ img_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/input_image.png?raw=true"
19
+ mask_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/mask.png?raw=true"
20
+ example_url = "https://github.com/IDEA-Research/detrex-storage/blob/main/assets/grounded_sam/paint_by_example/pomeranian_example.jpg?raw=True"
21
+ # example_url = "https://raw.githubusercontent.com/Fantasy-Studio/Paint-by-Example/main/examples/reference/example_1.jpg"
22
+
23
+ init_image = download_image(img_url).resize((512, 512))
24
+ mask_image = download_image(mask_url).resize((512, 512))
25
+ example_image = download_image(example_url).resize((512, 512))
26
+
27
+
28
+ """
29
+ Step 2: Download pretrained weights and initialize model
30
+ """
31
+ # set cache dir to store the weights
32
+ cache_dir = "/comp_robot/rentianhe/weights/diffusers/"
33
+
34
+ pipe = DiffusionPipeline.from_pretrained(
35
+ "Fantasy-Studio/Paint-by-Example",
36
+ torch_dtype=torch.float16,
37
+ cache_dir=cache_dir,
38
+ )
39
+ # set to device
40
+ pipe = pipe.to("cuda:1")
41
+
42
+
43
+ """
44
+ Step 3: Run PaintByExample pipeline and save image
45
+ """
46
+ image = pipe(
47
+ image=init_image,
48
+ mask_image=mask_image,
49
+ example_image=example_image,
50
+ num_inference_steps=200,
51
+ ).images[0]
52
+
53
+ image.save("./paint_by_example_demo.jpg")
external/Grounded-Segment-Anything/voxelnext_3d_box/models/__init__.py ADDED
File without changes
external/Grounded-Segment-Anything/voxelnext_3d_box/models/data_processor.py ADDED
@@ -0,0 +1,211 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from functools import partial
2
+
3
+ import numpy as np
4
+
5
+ tv = None
6
+ try:
7
+ import cumm.tensorview as tv
8
+ except:
9
+ pass
10
+
11
+
12
+ def mask_points_by_range(points, limit_range):
13
+ mask = (points[:, 0] >= limit_range[0]) & (points[:, 0] <= limit_range[3]) \
14
+ & (points[:, 1] >= limit_range[1]) & (points[:, 1] <= limit_range[4])
15
+ return mask
16
+
17
+
18
+ class VoxelGeneratorWrapper():
19
+ def __init__(self, vsize_xyz, coors_range_xyz, num_point_features, max_num_points_per_voxel, max_num_voxels):
20
+ try:
21
+ from spconv.utils import VoxelGeneratorV2 as VoxelGenerator
22
+ self.spconv_ver = 1
23
+ except:
24
+ try:
25
+ from spconv.utils import VoxelGenerator
26
+ self.spconv_ver = 1
27
+ except:
28
+ from spconv.utils import Point2VoxelCPU3d as VoxelGenerator
29
+ self.spconv_ver = 2
30
+
31
+ if self.spconv_ver == 1:
32
+ self._voxel_generator = VoxelGenerator(
33
+ voxel_size=vsize_xyz,
34
+ point_cloud_range=coors_range_xyz,
35
+ max_num_points=max_num_points_per_voxel,
36
+ max_voxels=max_num_voxels
37
+ )
38
+ else:
39
+ self._voxel_generator = VoxelGenerator(
40
+ vsize_xyz=vsize_xyz,
41
+ coors_range_xyz=coors_range_xyz,
42
+ num_point_features=num_point_features,
43
+ max_num_points_per_voxel=max_num_points_per_voxel,
44
+ max_num_voxels=max_num_voxels
45
+ )
46
+
47
+ def generate(self, points):
48
+ if self.spconv_ver == 1:
49
+ voxel_output = self._voxel_generator.generate(points)
50
+ if isinstance(voxel_output, dict):
51
+ voxels, coordinates, num_points = \
52
+ voxel_output['voxels'], voxel_output['coordinates'], voxel_output['num_points_per_voxel']
53
+ else:
54
+ voxels, coordinates, num_points = voxel_output
55
+ else:
56
+ assert tv is not None, f"Unexpected error, library: 'cumm' wasn't imported properly."
57
+ voxel_output = self._voxel_generator.point_to_voxel(tv.from_numpy(points))
58
+ tv_voxels, tv_coordinates, tv_num_points = voxel_output
59
+ # make copy with numpy(), since numpy_view() will disappear as soon as the generator is deleted
60
+ voxels = tv_voxels.numpy()
61
+ coordinates = tv_coordinates.numpy()
62
+ num_points = tv_num_points.numpy()
63
+ return voxels, coordinates, num_points
64
+
65
+
66
+ class DataProcessor(object):
67
+ def __init__(self, processor_configs, point_cloud_range, training, num_point_features):
68
+ self.point_cloud_range = point_cloud_range
69
+ self.training = training
70
+ self.num_point_features = num_point_features
71
+ self.mode = 'train' if training else 'test'
72
+ self.grid_size = self.voxel_size = None
73
+ self.data_processor_queue = []
74
+
75
+ self.voxel_generator = None
76
+
77
+ for cur_cfg in processor_configs:
78
+ cur_processor = getattr(self, cur_cfg.NAME)(config=cur_cfg)
79
+ self.data_processor_queue.append(cur_processor)
80
+
81
+ def mask_points_and_boxes_outside_range(self, data_dict=None, config=None):
82
+ if data_dict is None:
83
+ return partial(self.mask_points_and_boxes_outside_range, config=config)
84
+
85
+ if data_dict.get('points', None) is not None:
86
+ mask = mask_points_by_range(data_dict['points'], self.point_cloud_range)
87
+ data_dict['points'] = data_dict['points'][mask]
88
+
89
+ return data_dict
90
+
91
+ def shuffle_points(self, data_dict=None, config=None):
92
+ if data_dict is None:
93
+ return partial(self.shuffle_points, config=config)
94
+
95
+ if config.SHUFFLE_ENABLED[self.mode]:
96
+ points = data_dict['points']
97
+ shuffle_idx = np.random.permutation(points.shape[0])
98
+ points = points[shuffle_idx]
99
+ data_dict['points'] = points
100
+
101
+ return data_dict
102
+
103
+ def transform_points_to_voxels_placeholder(self, data_dict=None, config=None):
104
+ # just calculate grid size
105
+ if data_dict is None:
106
+ grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
107
+ self.grid_size = np.round(grid_size).astype(np.int64)
108
+ self.voxel_size = config.VOXEL_SIZE
109
+ return partial(self.transform_points_to_voxels_placeholder, config=config)
110
+
111
+ return data_dict
112
+
113
+ def double_flip(self, points):
114
+ # y flip
115
+ points_yflip = points.copy()
116
+ points_yflip[:, 1] = -points_yflip[:, 1]
117
+
118
+ # x flip
119
+ points_xflip = points.copy()
120
+ points_xflip[:, 0] = -points_xflip[:, 0]
121
+
122
+ # x y flip
123
+ points_xyflip = points.copy()
124
+ points_xyflip[:, 0] = -points_xyflip[:, 0]
125
+ points_xyflip[:, 1] = -points_xyflip[:, 1]
126
+
127
+ return points_yflip, points_xflip, points_xyflip
128
+
129
+ def transform_points_to_voxels(self, data_dict=None, config=None):
130
+ if data_dict is None:
131
+ grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
132
+ self.grid_size = np.round(grid_size).astype(np.int64)
133
+ self.voxel_size = config.VOXEL_SIZE
134
+ # just bind the config, we will create the VoxelGeneratorWrapper later,
135
+ # to avoid pickling issues in multiprocess spawn
136
+ return partial(self.transform_points_to_voxels, config=config)
137
+
138
+ if self.voxel_generator is None:
139
+ self.voxel_generator = VoxelGeneratorWrapper(
140
+ vsize_xyz=config.VOXEL_SIZE,
141
+ coors_range_xyz=self.point_cloud_range,
142
+ num_point_features=self.num_point_features,
143
+ max_num_points_per_voxel=config.MAX_POINTS_PER_VOXEL,
144
+ max_num_voxels=config.MAX_NUMBER_OF_VOXELS[self.mode],
145
+ )
146
+
147
+ points = data_dict['points']
148
+ voxel_output = self.voxel_generator.generate(points)
149
+ voxels, coordinates, num_points = voxel_output
150
+
151
+ data_dict['voxels'] = voxels
152
+ data_dict['voxel_coords'] = coordinates
153
+ data_dict['voxel_num_points'] = num_points
154
+ return data_dict
155
+
156
+ def sample_points(self, data_dict=None, config=None):
157
+ if data_dict is None:
158
+ return partial(self.sample_points, config=config)
159
+
160
+ num_points = config.NUM_POINTS[self.mode]
161
+ if num_points == -1:
162
+ return data_dict
163
+
164
+ points = data_dict['points']
165
+ if num_points < len(points):
166
+ pts_depth = np.linalg.norm(points[:, 0:3], axis=1)
167
+ pts_near_flag = pts_depth < 40.0
168
+ far_idxs_choice = np.where(pts_near_flag == 0)[0]
169
+ near_idxs = np.where(pts_near_flag == 1)[0]
170
+ choice = []
171
+ if num_points > len(far_idxs_choice):
172
+ near_idxs_choice = np.random.choice(near_idxs, num_points - len(far_idxs_choice), replace=False)
173
+ choice = np.concatenate((near_idxs_choice, far_idxs_choice), axis=0) \
174
+ if len(far_idxs_choice) > 0 else near_idxs_choice
175
+ else:
176
+ choice = np.arange(0, len(points), dtype=np.int32)
177
+ choice = np.random.choice(choice, num_points, replace=False)
178
+ np.random.shuffle(choice)
179
+ else:
180
+ choice = np.arange(0, len(points), dtype=np.int32)
181
+ if num_points > len(points):
182
+ extra_choice = np.random.choice(choice, num_points - len(points), replace=False)
183
+ choice = np.concatenate((choice, extra_choice), axis=0)
184
+ np.random.shuffle(choice)
185
+ data_dict['points'] = points[choice]
186
+ return data_dict
187
+
188
+ def calculate_grid_size(self, data_dict=None, config=None):
189
+ if data_dict is None:
190
+ grid_size = (self.point_cloud_range[3:6] - self.point_cloud_range[0:3]) / np.array(config.VOXEL_SIZE)
191
+ self.grid_size = np.round(grid_size).astype(np.int64)
192
+ self.voxel_size = config.VOXEL_SIZE
193
+ return partial(self.calculate_grid_size, config=config)
194
+ return data_dict
195
+
196
+ def forward(self, data_dict):
197
+ """
198
+ Args:
199
+ data_dict:
200
+ points: (N, 3 + C_in)
201
+ gt_boxes: optional, (N, 7 + C) [x, y, z, dx, dy, dz, heading, ...]
202
+ gt_names: optional, (N), string
203
+ ...
204
+
205
+ Returns:
206
+ """
207
+
208
+ for cur_processor in self.data_processor_queue:
209
+ data_dict = cur_processor(data_dict=data_dict)
210
+
211
+ return data_dict
external/Grounded-Segment-Anything/voxelnext_3d_box/models/mean_vfe.py ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ class MeanVFE(nn.Module):
5
+ def __init__(self):
6
+ super().__init__()
7
+
8
+ def forward(self, batch_dict, **kwargs):
9
+ """
10
+ Args:
11
+ batch_dict:
12
+ voxels: (num_voxels, max_points_per_voxel, C)
13
+ voxel_num_points: optional (num_voxels)
14
+ **kwargs:
15
+
16
+ Returns:
17
+ vfe_features: (num_voxels, C)
18
+ """
19
+ voxel_features, voxel_num_points = batch_dict['voxels'], batch_dict['voxel_num_points']
20
+ points_mean = voxel_features[:, :, :].sum(dim=1, keepdim=False)
21
+ normalizer = torch.clamp_min(voxel_num_points.view(-1, 1), min=1.0).type_as(voxel_features)
22
+ points_mean = points_mean / normalizer
23
+ batch_dict['voxel_features'] = points_mean.contiguous()
24
+
25
+ return batch_dict
external/Grounded-Segment-Anything/voxelnext_3d_box/models/spconv_backbone_voxelnext.py ADDED
@@ -0,0 +1,292 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from functools import partial
2
+ import torch
3
+ import torch.nn as nn
4
+
5
+ import spconv.pytorch as spconv
6
+ from spconv.core import ConvAlgo
7
+
8
+
9
+ def replace_feature(out, new_features):
10
+ return out.replace_feature(new_features)
11
+
12
+
13
+ def post_act_block(in_channels, out_channels, kernel_size, indice_key=None, stride=1, padding=0,
14
+ conv_type='subm', norm_fn=None):
15
+
16
+ if conv_type == 'subm':
17
+ conv = spconv.SubMConv3d(in_channels, out_channels, kernel_size, bias=False, indice_key=indice_key, algo=ConvAlgo.Native)
18
+ elif conv_type == 'spconv':
19
+ conv = spconv.SparseConv3d(in_channels, out_channels, kernel_size, stride=stride, padding=padding,
20
+ bias=False, indice_key=indice_key, algo=ConvAlgo.Native)
21
+ elif conv_type == 'inverseconv':
22
+ conv = spconv.SparseInverseConv3d(in_channels, out_channels, kernel_size, indice_key=indice_key, bias=False, algo=ConvAlgo.Native)
23
+ else:
24
+ raise NotImplementedError
25
+
26
+ m = spconv.SparseSequential(
27
+ conv,
28
+ norm_fn(out_channels),
29
+ nn.ReLU(),
30
+ )
31
+
32
+ return m
33
+
34
+
35
+ class SparseBasicBlock(spconv.SparseModule):
36
+ expansion = 1
37
+
38
+ def __init__(self, inplanes, planes, stride=1, norm_fn=None, downsample=None, indice_key=None):
39
+ super(SparseBasicBlock, self).__init__()
40
+
41
+ assert norm_fn is not None
42
+ bias = norm_fn is not None
43
+ self.conv1 = spconv.SubMConv3d(
44
+ inplanes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key, algo=ConvAlgo.Native
45
+ )
46
+ self.bn1 = norm_fn(planes)
47
+ self.relu = nn.ReLU()
48
+ self.conv2 = spconv.SubMConv3d(
49
+ planes, planes, kernel_size=3, stride=stride, padding=1, bias=bias, indice_key=indice_key, algo=ConvAlgo.Native
50
+ )
51
+ self.bn2 = norm_fn(planes)
52
+ self.downsample = downsample
53
+ self.stride = stride
54
+
55
+ def forward(self, x):
56
+ identity = x
57
+
58
+ out = self.conv1(x)
59
+ out = replace_feature(out, self.bn1(out.features))
60
+ out = replace_feature(out, self.relu(out.features))
61
+
62
+ out = self.conv2(out)
63
+ out = replace_feature(out, self.bn2(out.features))
64
+
65
+ if self.downsample is not None:
66
+ identity = self.downsample(x)
67
+
68
+ out = replace_feature(out, out.features + identity.features)
69
+ out = replace_feature(out, self.relu(out.features))
70
+
71
+ return out
72
+
73
+
74
+ class VoxelResBackBone8xVoxelNeXt(nn.Module):
75
+ def __init__(self, input_channels, grid_size, **kwargs):
76
+ super().__init__()
77
+ norm_fn = partial(nn.BatchNorm1d, eps=1e-3, momentum=0.01)
78
+
79
+ spconv_kernel_sizes = [3, 3, 3, 3]
80
+ channels = [16, 32, 64, 128, 128]
81
+ out_channel = 128
82
+
83
+ self.sparse_shape = grid_size[::-1] + [1, 0, 0]
84
+
85
+ self.conv_input = spconv.SparseSequential(
86
+ spconv.SubMConv3d(input_channels, channels[0], 3, padding=1, bias=False, indice_key='subm1', algo=ConvAlgo.Native),
87
+ norm_fn(channels[0]),
88
+ nn.ReLU(),
89
+ )
90
+ block = post_act_block
91
+
92
+ self.conv1 = spconv.SparseSequential(
93
+ SparseBasicBlock(channels[0], channels[0], norm_fn=norm_fn, indice_key='res1'),
94
+ SparseBasicBlock(channels[0], channels[0], norm_fn=norm_fn, indice_key='res1'),
95
+ )
96
+
97
+ self.conv2 = spconv.SparseSequential(
98
+ # [1600, 1408, 41] <- [800, 704, 21]
99
+ block(channels[0], channels[1], spconv_kernel_sizes[0], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[0]//2), indice_key='spconv2', conv_type='spconv'),
100
+ SparseBasicBlock(channels[1], channels[1], norm_fn=norm_fn, indice_key='res2'),
101
+ SparseBasicBlock(channels[1], channels[1], norm_fn=norm_fn, indice_key='res2'),
102
+ )
103
+
104
+ self.conv3 = spconv.SparseSequential(
105
+ # [800, 704, 21] <- [400, 352, 11]
106
+ block(channels[1], channels[2], spconv_kernel_sizes[1], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[1]//2), indice_key='spconv3', conv_type='spconv'),
107
+ SparseBasicBlock(channels[2], channels[2], norm_fn=norm_fn, indice_key='res3'),
108
+ SparseBasicBlock(channels[2], channels[2], norm_fn=norm_fn, indice_key='res3'),
109
+ )
110
+
111
+ self.conv4 = spconv.SparseSequential(
112
+ # [400, 352, 11] <- [200, 176, 6]
113
+ block(channels[2], channels[3], spconv_kernel_sizes[2], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[2]//2), indice_key='spconv4', conv_type='spconv'),
114
+ SparseBasicBlock(channels[3], channels[3], norm_fn=norm_fn, indice_key='res4'),
115
+ SparseBasicBlock(channels[3], channels[3], norm_fn=norm_fn, indice_key='res4'),
116
+ )
117
+
118
+ self.conv5 = spconv.SparseSequential(
119
+ # [200, 176, 6] <- [100, 88, 3]
120
+ block(channels[3], channels[4], spconv_kernel_sizes[3], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[3]//2), indice_key='spconv5', conv_type='spconv'),
121
+ SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res5'),
122
+ SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res5'),
123
+ )
124
+
125
+ self.conv6 = spconv.SparseSequential(
126
+ # [200, 176, 6] <- [100, 88, 3]
127
+ block(channels[4], channels[4], spconv_kernel_sizes[3], norm_fn=norm_fn, stride=2, padding=int(spconv_kernel_sizes[3]//2), indice_key='spconv6', conv_type='spconv'),
128
+ SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res6'),
129
+ SparseBasicBlock(channels[4], channels[4], norm_fn=norm_fn, indice_key='res6'),
130
+ )
131
+ self.conv_out = spconv.SparseSequential(
132
+ # [200, 150, 5] -> [200, 150, 2]
133
+ spconv.SparseConv2d(channels[3], out_channel, 3, stride=1, padding=1, bias=False, indice_key='spconv_down2', algo=ConvAlgo.Native),
134
+ norm_fn(out_channel),
135
+ nn.ReLU(),
136
+ )
137
+
138
+ self.shared_conv = spconv.SparseSequential(
139
+ spconv.SubMConv2d(out_channel, out_channel, 3, stride=1, padding=1, bias=True, algo=ConvAlgo.Native),
140
+ nn.BatchNorm1d(out_channel),
141
+ nn.ReLU(True),
142
+ )
143
+
144
+ self.forward_ret_dict = {}
145
+ self.num_point_features = out_channel
146
+ self.backbone_channels = {
147
+ 'x_conv1': channels[0],
148
+ 'x_conv2': channels[1],
149
+ 'x_conv3': channels[2],
150
+ 'x_conv4': channels[3]
151
+ }
152
+
153
+ def bev_out(self, x_conv, index):
154
+ features_cat = x_conv.features
155
+ indices_cat = x_conv.indices[:, [0, 2, 3]]
156
+ spatial_shape = x_conv.spatial_shape[1:]
157
+
158
+ indices_unique, _inv = torch.unique(indices_cat, dim=0, return_inverse=True)
159
+ features_unique = features_cat.new_zeros((indices_unique.shape[0], features_cat.shape[1]))
160
+ features_unique.index_add_(0, _inv, features_cat)
161
+
162
+ perm = torch.arange(_inv.size(0), dtype=_inv.dtype, device=_inv.device)
163
+ perm = _inv.new_empty(indices_unique.size(0)).scatter_(0, _inv, perm)
164
+ index_out = index[perm]
165
+
166
+ x_out = spconv.SparseConvTensor(
167
+ features=features_unique,
168
+ indices=indices_unique,
169
+ spatial_shape=spatial_shape,
170
+ batch_size=x_conv.batch_size
171
+ )
172
+ return x_out, index_out
173
+
174
+ def track_voxels_2d(self, x, x_downsample, index, kernel_size=3):
175
+ _step = int(kernel_size//2)
176
+ kernel_offsets = [[i, j] for i in range(-_step, _step+1) for j in range(-_step, _step+1)]
177
+ #kernel_offsets.remove([0, 0])
178
+ kernel_offsets = torch.Tensor(kernel_offsets).to(x.indices.device)
179
+
180
+ batch_size = x.batch_size
181
+ index_batch = []
182
+ indices_batch = []
183
+
184
+ for b in range(batch_size):
185
+ batch_index = x.indices[:, 0]==b
186
+ indices_ori = x.indices[batch_index]
187
+ features_ori = index[batch_index]
188
+
189
+ features_fore = features_ori
190
+ coords_fore = indices_ori
191
+
192
+ voxel_kerels_imp = kernel_offsets.unsqueeze(0).repeat(features_fore.shape[0],1, 1)
193
+ indices_fore_kernels = coords_fore[:, 1:].unsqueeze(1).repeat(1, kernel_offsets.shape[0], 1)
194
+ indices_with_imp = indices_fore_kernels + voxel_kerels_imp
195
+ features_fore = features_fore.repeat(1, kernel_offsets.shape[0])
196
+
197
+ selected_indices = indices_with_imp
198
+ spatial_indices = (selected_indices[:, :, 0] >=0) * (selected_indices[:, :, 1] >=0) * \
199
+ (selected_indices[:, :, 0] < x.spatial_shape[0]) * (selected_indices[:, :, 1] < x.spatial_shape[1])
200
+ selected_indices = selected_indices[spatial_indices]
201
+ features_fore = features_fore[spatial_indices].view(-1, 1)
202
+
203
+ selected_indices = torch.cat([torch.ones((selected_indices.shape[0], 1), device=features_fore.device)*b, selected_indices], dim=1)
204
+
205
+ features_fore, coords_fore = features_fore, selected_indices
206
+ index_batch.append(features_fore)
207
+ indices_batch.append(coords_fore)
208
+
209
+ index_batch = torch.cat(index_batch)
210
+ indices_batch = torch.cat(indices_batch)
211
+
212
+ return self.index_from_sparse(index_batch, indices_batch, x_downsample, True)
213
+
214
+ def index_from_sparse(self, feature, indices, x_target, _2d=False):
215
+ sparse_index = spconv.SparseConvTensor(
216
+ features=feature,
217
+ indices=indices.int(),
218
+ spatial_shape=x_target.spatial_shape,
219
+ batch_size=x_target.batch_size
220
+ )
221
+ dense_index = sparse_index.dense()
222
+ indices_downsample = x_target.indices.long()
223
+ if _2d:
224
+ index_downsample = dense_index[indices_downsample[:, 0], :, indices_downsample[:, 1], indices_downsample[:, 2]]
225
+ else:
226
+ index_downsample = dense_index[indices_downsample[:, 0], :, indices_downsample[:, 1], indices_downsample[:, 2], indices_downsample[:, 3]]
227
+ return index_downsample
228
+
229
+ def forward(self, batch_dict):
230
+ """
231
+ Args:
232
+ batch_dict:
233
+ batch_size: int
234
+ vfe_features: (num_voxels, C)
235
+ voxel_coords: (num_voxels, 4), [batch_idx, z_idx, y_idx, x_idx]
236
+ Returns:
237
+ batch_dict:
238
+ encoded_spconv_tensor: sparse tensor
239
+ """
240
+ voxel_features, voxel_coords = batch_dict['voxel_features'], batch_dict['voxel_coords']
241
+ batch_size = batch_dict['batch_size']
242
+ input_sp_tensor = spconv.SparseConvTensor(
243
+ features=voxel_features,
244
+ indices=voxel_coords.int(),
245
+ spatial_shape=self.sparse_shape,
246
+ batch_size=batch_size
247
+ )
248
+ x = self.conv_input(input_sp_tensor)
249
+
250
+ x_conv1 = self.conv1(x)
251
+ x_conv2 = self.conv2(x_conv1)
252
+ x_conv3 = self.conv3(x_conv2)
253
+ x_conv4 = self.conv4(x_conv3)
254
+ x_conv5 = self.conv5(x_conv4)
255
+ x_conv6 = self.conv6(x_conv5)
256
+
257
+ x_conv5.indices[:, 1:] *= 2
258
+ x_conv6.indices[:, 1:] *= 4
259
+ x_conv4 = x_conv4.replace_feature(torch.cat([x_conv4.features, x_conv5.features, x_conv6.features]))
260
+ x_conv4.indices = torch.cat([x_conv4.indices, x_conv5.indices, x_conv6.indices])
261
+
262
+ index6_out = torch.arange(x_conv4.indices.shape[0], device=x_conv4.indices.device).unsqueeze(-1)
263
+ out_bevout, index_bevout = self.bev_out(x_conv4, index6_out)
264
+
265
+ out = self.conv_out(out_bevout)
266
+ index_out = self.track_voxels_2d(out_bevout, out, index_bevout)
267
+
268
+ out = self.shared_conv(out)
269
+
270
+ batch_dict.update({
271
+ 'encoded_spconv_tensor': out,
272
+ 'encoded_spconv_tensor_stride': 8,
273
+ 'out_voxels': x_conv4.indices[index_out.squeeze(-1)],
274
+ })
275
+ batch_dict.update({
276
+ 'multi_scale_3d_features': {
277
+ 'x_conv1': x_conv1,
278
+ 'x_conv2': x_conv2,
279
+ 'x_conv3': x_conv3,
280
+ 'x_conv4': x_conv4,
281
+ }
282
+ })
283
+ batch_dict.update({
284
+ 'multi_scale_3d_strides': {
285
+ 'x_conv1': 1,
286
+ 'x_conv2': 2,
287
+ 'x_conv3': 4,
288
+ 'x_conv4': 8,
289
+ }
290
+ })
291
+
292
+ return batch_dict
external/Grounded-Segment-Anything/voxelnext_3d_box/models/voxelnext_head.py ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import numpy as np
2
+ import torch
3
+ import torch.nn as nn
4
+ from voxelnext_3d_box.utils import centernet_utils
5
+ import spconv.pytorch as spconv
6
+ import copy
7
+ from spconv.core import ConvAlgo
8
+
9
+
10
+ class SeparateHead(nn.Module):
11
+ def __init__(self, input_channels, sep_head_dict, kernel_size, use_bias=False):
12
+ super().__init__()
13
+ self.sep_head_dict = sep_head_dict
14
+
15
+ for cur_name in self.sep_head_dict:
16
+ output_channels = self.sep_head_dict[cur_name]['out_channels']
17
+ num_conv = self.sep_head_dict[cur_name]['num_conv']
18
+
19
+ fc_list = []
20
+ for k in range(num_conv - 1):
21
+ fc_list.append(spconv.SparseSequential(
22
+ spconv.SubMConv2d(input_channels, input_channels, kernel_size, padding=int(kernel_size//2), bias=use_bias, indice_key=cur_name, algo=ConvAlgo.Native),
23
+ nn.BatchNorm1d(input_channels),
24
+ nn.ReLU()
25
+ ))
26
+ fc_list.append(spconv.SubMConv2d(input_channels, output_channels, 1, bias=True, indice_key=cur_name+'out', algo=ConvAlgo.Native))
27
+ fc = nn.Sequential(*fc_list)
28
+ self.__setattr__(cur_name, fc)
29
+
30
+ def forward(self, x):
31
+ ret_dict = {}
32
+ for cur_name in self.sep_head_dict:
33
+ ret_dict[cur_name] = self.__getattr__(cur_name)(x).features
34
+
35
+ return ret_dict
36
+
37
+
38
+ class VoxelNeXtHead(nn.Module):
39
+ def __init__(self, class_names, point_cloud_range, voxel_size, kernel_size_head,
40
+ CLASS_NAMES_EACH_HEAD, SEPARATE_HEAD_CFG, POST_PROCESSING):
41
+ super().__init__()
42
+ self.point_cloud_range = torch.Tensor(point_cloud_range)
43
+ self.voxel_size = torch.Tensor(voxel_size)
44
+ self.feature_map_stride = 8
45
+
46
+ self.class_names = class_names
47
+ self.class_names_each_head = []
48
+ self.class_id_mapping_each_head = []
49
+ self.POST_PROCESSING = POST_PROCESSING
50
+
51
+ for cur_class_names in CLASS_NAMES_EACH_HEAD:
52
+ self.class_names_each_head.append([x for x in cur_class_names if x in class_names])
53
+ cur_class_id_mapping = torch.from_numpy(np.array(
54
+ [self.class_names.index(x) for x in cur_class_names if x in class_names]
55
+ ))
56
+ self.class_id_mapping_each_head.append(cur_class_id_mapping)
57
+
58
+ total_classes = sum([len(x) for x in self.class_names_each_head])
59
+ assert total_classes == len(self.class_names), f'class_names_each_head={self.class_names_each_head}'
60
+
61
+ self.heads_list = nn.ModuleList()
62
+ self.separate_head_cfg = SEPARATE_HEAD_CFG
63
+ for idx, cur_class_names in enumerate(self.class_names_each_head):
64
+ cur_head_dict = copy.deepcopy(self.separate_head_cfg.HEAD_DICT)
65
+ cur_head_dict['hm'] = dict(out_channels=len(cur_class_names), num_conv=2)
66
+ self.heads_list.append(
67
+ SeparateHead(
68
+ input_channels=128,
69
+ sep_head_dict=cur_head_dict,
70
+ kernel_size=kernel_size_head,
71
+ use_bias=True,
72
+ )
73
+ )
74
+ self.forward_ret_dict = {}
75
+
76
+ def generate_predicted_boxes(self, batch_size, pred_dicts, voxel_indices, spatial_shape):
77
+ device = pred_dicts[0]['hm'].device
78
+ post_process_cfg = self.POST_PROCESSING
79
+ post_center_limit_range = torch.tensor(post_process_cfg.POST_CENTER_LIMIT_RANGE).float().to(device)
80
+
81
+ ret_dict = [{
82
+ 'pred_boxes': [],
83
+ 'pred_scores': [],
84
+ 'pred_labels': [],
85
+ 'pred_ious': [],
86
+ 'voxel_ids': []
87
+ } for k in range(batch_size)]
88
+ for idx, pred_dict in enumerate(pred_dicts):
89
+ batch_hm = pred_dict['hm'].sigmoid()
90
+ batch_center = pred_dict['center']
91
+ batch_center_z = pred_dict['center_z']
92
+ batch_dim = pred_dict['dim'].exp()
93
+ batch_rot_cos = pred_dict['rot'][:, 0].unsqueeze(dim=1)
94
+ batch_rot_sin = pred_dict['rot'][:, 1].unsqueeze(dim=1)
95
+ batch_iou = None
96
+ batch_vel = pred_dict['vel'] if 'vel' in self.separate_head_cfg.HEAD_ORDER else None
97
+ voxel_indices_ = voxel_indices
98
+
99
+ final_pred_dicts = centernet_utils.decode_bbox_from_voxels_nuscenes(
100
+ batch_size=batch_size, indices=voxel_indices_,
101
+ obj=batch_hm,
102
+ rot_cos=batch_rot_cos,
103
+ rot_sin=batch_rot_sin,
104
+ center=batch_center, center_z=batch_center_z,
105
+ dim=batch_dim, vel=batch_vel, iou=batch_iou,
106
+ point_cloud_range=self.point_cloud_range.to(device), voxel_size=self.voxel_size.to(device),
107
+ feature_map_stride=self.feature_map_stride,
108
+ K=post_process_cfg.MAX_OBJ_PER_SAMPLE,
109
+ score_thresh=post_process_cfg.SCORE_THRESH,
110
+ post_center_limit_range=post_center_limit_range,
111
+ add_features=torch.arange(voxel_indices_.shape[0], device=voxel_indices_.device).unsqueeze(-1)
112
+ )
113
+
114
+ for k, final_dict in enumerate(final_pred_dicts):
115
+ class_id_mapping_each_head = self.class_id_mapping_each_head[idx].to(device)
116
+ final_dict['pred_labels'] = class_id_mapping_each_head[final_dict['pred_labels'].long()]
117
+
118
+ ret_dict[k]['pred_boxes'].append(final_dict['pred_boxes'])
119
+ ret_dict[k]['pred_scores'].append(final_dict['pred_scores'])
120
+ ret_dict[k]['pred_labels'].append(final_dict['pred_labels'])
121
+ ret_dict[k]['pred_ious'].append(final_dict['pred_ious'])
122
+ ret_dict[k]['voxel_ids'].append(final_dict['add_features'])
123
+
124
+ for k in range(batch_size):
125
+ pred_boxes = torch.cat(ret_dict[k]['pred_boxes'], dim=0)
126
+ pred_scores = torch.cat(ret_dict[k]['pred_scores'], dim=0)
127
+ pred_labels = torch.cat(ret_dict[k]['pred_labels'], dim=0)
128
+ voxel_ids = torch.cat(ret_dict[k]['voxel_ids'], dim=0)
129
+
130
+ ret_dict[k]['pred_boxes'] = pred_boxes
131
+ ret_dict[k]['pred_scores'] = pred_scores
132
+ ret_dict[k]['pred_labels'] = pred_labels + 1
133
+ ret_dict[k]['voxel_ids'] = voxel_ids
134
+
135
+ return ret_dict
136
+
137
+ def _get_voxel_infos(self, x):
138
+ spatial_shape = x.spatial_shape
139
+ voxel_indices = x.indices
140
+ spatial_indices = []
141
+ num_voxels = []
142
+ batch_size = x.batch_size
143
+ batch_index = voxel_indices[:, 0]
144
+
145
+ for bs_idx in range(batch_size):
146
+ batch_inds = batch_index==bs_idx
147
+ spatial_indices.append(voxel_indices[batch_inds][:, [2, 1]])
148
+ num_voxels.append(batch_inds.sum())
149
+
150
+ return spatial_shape, batch_index, voxel_indices, spatial_indices, num_voxels
151
+
152
+ def forward(self, data_dict):
153
+ x = data_dict['encoded_spconv_tensor']
154
+ spatial_shape, batch_index, voxel_indices, spatial_indices, num_voxels = self._get_voxel_infos(x)
155
+
156
+ pred_dicts = []
157
+ for idx, head in enumerate(self.heads_list):
158
+ pred_dict = head(x)
159
+ pred_dicts.append(pred_dict)
160
+
161
+ pred_dicts = self.generate_predicted_boxes(
162
+ data_dict['batch_size'],
163
+ pred_dicts, voxel_indices, spatial_shape
164
+ )
165
+
166
+ return pred_dicts
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/centernet_utils.py ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # This file is modified from https://github.com/tianweiy/CenterPoint
2
+
3
+ import torch
4
+
5
+
6
+ def _topk_1d(scores, batch_size, batch_idx, obj, K=40, nuscenes=False):
7
+ # scores: (N, num_classes)
8
+ topk_score_list = []
9
+ topk_inds_list = []
10
+ topk_classes_list = []
11
+
12
+ for bs_idx in range(batch_size):
13
+ batch_inds = batch_idx==bs_idx
14
+ if obj.shape[-1] == 1 and not nuscenes:
15
+ score = scores[batch_inds].permute(1, 0)
16
+ topk_scores, topk_inds = torch.topk(score, K)
17
+ topk_score, topk_ind = torch.topk(obj[topk_inds.view(-1)].squeeze(-1), K) #torch.topk(topk_scores.view(-1), K)
18
+ else:
19
+ score = obj[batch_inds].permute(1, 0)
20
+ topk_scores, topk_inds = torch.topk(score, min(K, score.shape[-1]))
21
+ topk_score, topk_ind = torch.topk(topk_scores.view(-1), min(K, topk_scores.view(-1).shape[-1]))
22
+
23
+ topk_classes = (topk_ind // K).int()
24
+ topk_inds = topk_inds.view(-1).gather(0, topk_ind)
25
+ #print('topk_inds', topk_inds)
26
+
27
+ if not obj is None and obj.shape[-1] == 1:
28
+ topk_score_list.append(obj[batch_inds][topk_inds])
29
+ else:
30
+ topk_score_list.append(topk_score)
31
+ topk_inds_list.append(topk_inds)
32
+ topk_classes_list.append(topk_classes)
33
+
34
+ topk_score = torch.stack(topk_score_list)
35
+ topk_inds = torch.stack(topk_inds_list)
36
+ topk_classes = torch.stack(topk_classes_list)
37
+
38
+ return topk_score, topk_inds, topk_classes
39
+
40
+ def gather_feat_idx(feats, inds, batch_size, batch_idx):
41
+ feats_list = []
42
+ dim = feats.size(-1)
43
+ _inds = inds.unsqueeze(-1).expand(inds.size(0), inds.size(1), dim)
44
+
45
+ for bs_idx in range(batch_size):
46
+ batch_inds = batch_idx==bs_idx
47
+ feat = feats[batch_inds]
48
+ feats_list.append(feat.gather(0, _inds[bs_idx]))
49
+ feats = torch.stack(feats_list)
50
+ return feats
51
+
52
+
53
+ def decode_bbox_from_voxels_nuscenes(batch_size, indices, obj, rot_cos, rot_sin,
54
+ center, center_z, dim, vel=None, iou=None, point_cloud_range=None, voxel_size=None, voxels_3d=None,
55
+ feature_map_stride=None, K=100, score_thresh=None, post_center_limit_range=None, add_features=None):
56
+ batch_idx = indices[:, 0]
57
+ spatial_indices = indices[:, 1:]
58
+ scores, inds, class_ids = _topk_1d(None, batch_size, batch_idx, obj, K=K, nuscenes=True)
59
+
60
+ center = gather_feat_idx(center, inds, batch_size, batch_idx)
61
+ rot_sin = gather_feat_idx(rot_sin, inds, batch_size, batch_idx)
62
+ rot_cos = gather_feat_idx(rot_cos, inds, batch_size, batch_idx)
63
+ center_z = gather_feat_idx(center_z, inds, batch_size, batch_idx)
64
+ dim = gather_feat_idx(dim, inds, batch_size, batch_idx)
65
+ spatial_indices = gather_feat_idx(spatial_indices, inds, batch_size, batch_idx)
66
+
67
+ if not add_features is None:
68
+ add_features = gather_feat_idx(add_features, inds, batch_size, batch_idx) #for add_feature in add_features]
69
+
70
+ if not isinstance(feature_map_stride, int):
71
+ feature_map_stride = gather_feat_idx(feature_map_stride.unsqueeze(-1), inds, batch_size, batch_idx)
72
+
73
+ angle = torch.atan2(rot_sin, rot_cos)
74
+ xs = (spatial_indices[:, :, -1:] + center[:, :, 0:1]) * feature_map_stride * voxel_size[0] + point_cloud_range[0]
75
+ ys = (spatial_indices[:, :, -2:-1] + center[:, :, 1:2]) * feature_map_stride * voxel_size[1] + point_cloud_range[1]
76
+
77
+ box_part_list = [xs, ys, center_z, dim, angle]
78
+
79
+ if not vel is None:
80
+ vel = gather_feat_idx(vel, inds, batch_size, batch_idx)
81
+ box_part_list.append(vel)
82
+
83
+ if not iou is None:
84
+ iou = gather_feat_idx(iou, inds, batch_size, batch_idx)
85
+ iou = torch.clamp(iou, min=0, max=1.)
86
+
87
+ final_box_preds = torch.cat((box_part_list), dim=-1)
88
+ final_scores = scores.view(batch_size, K)
89
+ final_class_ids = class_ids.view(batch_size, K)
90
+ if not add_features is None:
91
+ add_features = add_features.view(batch_size, K, add_features.shape[-1]) #for add_feature in add_features]
92
+
93
+ assert post_center_limit_range is not None
94
+ mask = (final_box_preds[..., :3] >= post_center_limit_range[:3]).all(2)
95
+ mask &= (final_box_preds[..., :3] <= post_center_limit_range[3:]).all(2)
96
+
97
+ if score_thresh is not None:
98
+ mask &= (final_scores > score_thresh)
99
+
100
+ ret_pred_dicts = []
101
+ for k in range(batch_size):
102
+ cur_mask = mask[k]
103
+ cur_boxes = final_box_preds[k, cur_mask]
104
+ cur_scores = final_scores[k, cur_mask]
105
+ cur_labels = final_class_ids[k, cur_mask]
106
+ cur_add_features = add_features[k, cur_mask] if not add_features is None else None
107
+ cur_iou = iou[k, cur_mask] if not iou is None else None
108
+
109
+ ret_pred_dicts.append({
110
+ 'pred_boxes': cur_boxes,
111
+ 'pred_scores': cur_scores,
112
+ 'pred_labels': cur_labels,
113
+ 'pred_ious': cur_iou,
114
+ 'add_features': cur_add_features,
115
+ })
116
+ return ret_pred_dicts
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/config.py ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import yaml
2
+ from easydict import EasyDict
3
+
4
+ def merge_new_config(config, new_config):
5
+ if '_BASE_CONFIG_' in new_config:
6
+ with open(new_config['_BASE_CONFIG_'], 'r') as f:
7
+ try:
8
+ yaml_config = yaml.safe_load(f, Loader=yaml.FullLoader)
9
+ except:
10
+ yaml_config = yaml.safe_load(f)
11
+ config.update(EasyDict(yaml_config))
12
+
13
+ for key, val in new_config.items():
14
+ if not isinstance(val, dict):
15
+ config[key] = val
16
+ continue
17
+ if key not in config:
18
+ config[key] = EasyDict()
19
+ merge_new_config(config[key], val)
20
+
21
+ return config
22
+
23
+ def cfg_from_yaml_file(cfg_file, config):
24
+ with open(cfg_file, 'r') as f:
25
+ try:
26
+ new_config = yaml.safe_load(f, Loader=yaml.FullLoader)
27
+ except:
28
+ new_config = yaml.safe_load(f)
29
+
30
+ merge_new_config(config=config, new_config=new_config)
31
+
32
+ return config
33
+
34
+ cfg = EasyDict()
external/Grounded-Segment-Anything/voxelnext_3d_box/utils/image_projection.py ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import numpy as np
3
+ import cv2
4
+
5
+ def get_data_info(info, cam_type):
6
+
7
+ cam_info = info[cam_type]
8
+
9
+ lidar2cam_r = np.linalg.inv(cam_info['sensor2lidar_rotation'])
10
+ lidar2cam_t = cam_info[
11
+ 'sensor2lidar_translation'] @ lidar2cam_r.T
12
+ lidar2cam_rt = np.eye(4)
13
+ lidar2cam_rt[:3, :3] = lidar2cam_r.T
14
+ lidar2cam_rt[3, :3] = -lidar2cam_t
15
+ intrinsic = cam_info['cam_intrinsic']
16
+ viewpad = np.eye(4)
17
+ viewpad[:intrinsic.shape[0], :intrinsic.shape[1]] = intrinsic
18
+ lidar2img_rt = (viewpad @ lidar2cam_rt.T)
19
+
20
+ return lidar2img_rt
21
+
22
+
23
+ def _proj_voxel_image(voxel_coords, lidar2img_rt, voxel_size, point_cloud_range):
24
+ # voxel_coords [n ,4]
25
+ # lidar2img_rt [4, 4]
26
+ #x_input.indices [n, 4] [[0, Z, Y, Z]xn]
27
+
28
+ voxel_coords = voxel_coords[:, [3,2,1]]
29
+ device = voxel_coords.device
30
+ lidar2img_rt = torch.Tensor(lidar2img_rt).to(device)
31
+ # point_cloud_rangetensor([-51.2000, -51.2000, -5.0000, 51.2000, 51.2000, 3.0000])
32
+ voxel_coords = voxel_coords * voxel_size.unsqueeze(0) + point_cloud_range[:3].unsqueeze(0)
33
+ # (n, 4)
34
+ voxel_coords = torch.cat([voxel_coords, torch.ones((voxel_coords.shape[0], 1), device=device)], dim=-1)
35
+ points_image = torch.matmul(lidar2img_rt, voxel_coords.permute(1, 0)) #(voxel_coords @ lidar2img_rt).T
36
+ # (4, n)
37
+ depth = points_image[2:3] # (1, n)
38
+ points_image = points_image[:2] / torch.maximum(depth, torch.ones_like(depth*1e-4))
39
+ return points_image, depth
40
+
41
+ def _draw_image(points_image, image_path, depth):
42
+ image = cv2.imread(image_path)
43
+ points_image = points_image.int().cpu().numpy()
44
+ for i in range(points_image.shape[1]):
45
+ _point = points_image[:, i]
46
+ if _point[0] > 0 and _point[1] > 0 and depth[0][i] >0:
47
+ cv2.circle(image, tuple(_point), 1, (0,255,0), -1)
48
+ #cv2.imwrite("image.png", image)
49
+ return image
50
+
51
+ def _draw_mask(image_path, mask, color=None):
52
+ image = cv2.imread(image_path)
53
+ h, w, _ = image.shape
54
+
55
+ if color is None:
56
+ color = np.random.random(3)
57
+
58
+ image[mask] = image[mask] * color
59
+ #cv2.imwrite("image_mask.png", image)
60
+ return image
61
+
62
+ def _draw_3dbox(box, lidar2img_rt, image, mask=None, color=None, output_path="image_box.png"):
63
+ #image = cv2.imread(image_path)
64
+ h, w, _ = image.shape
65
+
66
+ if color is None:
67
+ color = np.random.random(3)
68
+ if not mask is None:
69
+ image[mask] = image[mask] * color
70
+
71
+ center_x, center_y, center_z, H, W, Z, angle = box[:7]
72
+ sin_angle, cos_angle = torch.sin(angle), torch.cos(angle)
73
+ top1 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
74
+ top2 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
75
+ top3 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
76
+ top4 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z + Z/2]
77
+
78
+ down1 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
79
+ down2 = [center_x - (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
80
+ down3 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y + (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
81
+ down4 = [center_x + (H/2 * cos_angle + W/2 * sin_angle), center_y - (H/2 * sin_angle + W/2 * cos_angle), center_z - Z/2]
82
+ points = torch.Tensor([top1, top2, top3, top4, down1, down2, down3, down4, [center_x, center_y, center_z]]) # (8, 3)
83
+ points = torch.cat([points, torch.ones((points.shape[0], 1), device=points.device)], dim=-1)
84
+ points_image = torch.matmul(torch.Tensor(lidar2img_rt).to(points.device), points.permute(1, 0))
85
+ depth = points_image[2:3] # (1, n)
86
+ points_image = points_image[:2] / torch.maximum(depth, torch.ones_like(depth*1e-4))
87
+ points_image = points_image.permute(1, 0).int().cpu().numpy() #(voxel_coords @ lidar2img_rt).T
88
+ lines = [[0, 1], [1, 2], [2, 3], [3, 0], [4, 5], [5, 6], [6, 7], [7, 4], [0, 4], [1, 5], [2, 6], [3, 7]]
89
+
90
+ cv2.circle(image, tuple(points_image[-1]), 3, (0, 255, 0), -1)
91
+
92
+ for line in lines:
93
+ cv2.line(image, tuple(points_image[line[0]]), tuple(points_image[line[1]]), tuple(color * 255), 2)
94
+ #cv2.imwrite(output_path, image)
95
+ return image
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
external/WildCamera/WildCamera/benchmark/benchmark_calibration.py ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import print_function, division
2
+ import os, sys, inspect
3
+ project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
4
+ sys.path.insert(0, project_root)
5
+
6
+ import warnings
7
+ warnings.filterwarnings('ignore')
8
+ import torch
9
+ import argparse
10
+ from loguru import logger
11
+ from pprint import pprint
12
+
13
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
14
+
15
+
16
+ parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
17
+ parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
18
+ parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
19
+ parser.add_argument('--experiment_name', type=str, help='name of the experiment', required=True, choices=['in_the_wild', 'gsv'])
20
+
21
+ def main_worker(args, wtassumption=False):
22
+ args.gpu = 0
23
+ pprint(vars(args))
24
+
25
+ model = NEWCRFIF(version='large07', pretrained=None)
26
+ model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
27
+ model.eval()
28
+ model.cuda()
29
+ logger.info("Load Model from %s" % args.load_ckpt)
30
+
31
+ if args.experiment_name == 'in_the_wild':
32
+ from WildCamera.evaluation.evaluate_intrinsic import EvaluateIntrinsic
33
+ evaluate_intrinsic = EvaluateIntrinsic()
34
+ evaluate_intrinsic.evaluate(model, args, steps=0, writer=None, group=None, wtassumption=wtassumption)
35
+ elif args.experiment_name == 'gsv':
36
+ from WildCamera.evaluation.evaluate_fov import EvaluateFov
37
+ evaluate_fov = EvaluateFov()
38
+ evaluate_fov.evaluate(model, args, steps=0, writer=None, group=None, wtassumption=wtassumption)
39
+
40
+
41
+ if __name__ == '__main__':
42
+ args = parser.parse_args()
43
+
44
+ args.world_size = torch.cuda.device_count()
45
+ args.augscale = 2.0
46
+ args.input_height = 480
47
+ args.input_width = 640
48
+
49
+ args.eval_workers = 4
50
+ args.l1_th = 0.02
51
+
52
+ if args.experiment_name == 'in_the_wild':
53
+ args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_all.pth')
54
+ args.datasets_train = [
55
+ 'Nuscenes',
56
+ 'KITTI',
57
+ 'Cityscapes',
58
+ 'NYUv2',
59
+ 'ARKitScenes',
60
+ 'MegaDepth',
61
+ 'SUN3D',
62
+ 'MVImgNet',
63
+ 'Objectron',
64
+ ]
65
+ args.datasets_eval = [
66
+ 'Nuscenes',
67
+ 'KITTI',
68
+ 'Cityscapes',
69
+ 'NYUv2',
70
+ 'ARKitScenes',
71
+ 'MegaDepth',
72
+ 'SUN3D',
73
+ 'MVImgNet',
74
+ 'Objectron',
75
+ 'Waymo',
76
+ 'BIWIRGBDID',
77
+ 'RGBD',
78
+ 'ScanNet',
79
+ 'CAD120',
80
+ 'MVS',
81
+ 'Scenes11',
82
+ ]
83
+ main_worker(args, wtassumption=False)
84
+ main_worker(args, wtassumption=True)
85
+ elif args.experiment_name == 'gsv':
86
+ args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_gsv.pth')
87
+ args.datasets_train = [
88
+ 'GSV'
89
+ ]
90
+ args.datasets_eval = [
91
+ 'GSV'
92
+ ]
93
+ main_worker(args, wtassumption=False)
94
+ main_worker(args, wtassumption=True)
95
+ else:
96
+ raise NotImplementedError()
97
+
98
+
external/WildCamera/WildCamera/benchmark/benchmark_crop.py ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import print_function, division
2
+ import os, sys, inspect, copy
3
+ project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
4
+ sys.path.insert(0, project_root)
5
+
6
+ import warnings
7
+ warnings.filterwarnings('ignore')
8
+ import torch
9
+ import argparse
10
+ from loguru import logger
11
+ from pprint import pprint
12
+
13
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
14
+ from WildCamera.evaluation.evaluate_crop import EvaluateCrop
15
+
16
+ parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
17
+ parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
18
+ parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
19
+
20
+ def main_worker(args):
21
+ args.gpu = 0
22
+ pprint(vars(args))
23
+
24
+ model = NEWCRFIF(version='large07', pretrained=None)
25
+ model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
26
+ model.eval()
27
+ model.cuda()
28
+ logger.info("Load Model from %s" % args.load_ckpt)
29
+
30
+ evaluate_crop = EvaluateCrop()
31
+ evaluate_crop.evaluate(model, args, group=None)
32
+
33
+ if __name__ == '__main__':
34
+ args = parser.parse_args()
35
+
36
+ args.world_size = 1
37
+ args.augscale = 2.0
38
+ args.input_height = 480
39
+ args.input_width = 640
40
+ args.eval_workers = 4
41
+ args.l1_th = 0.02
42
+
43
+ args.datasets_train = [
44
+ 'KITTI',
45
+ 'NYUv2',
46
+ 'ARKitScenes',
47
+ 'Waymo',
48
+ 'RGBD',
49
+ 'ScanNet',
50
+ 'MVS',
51
+ ]
52
+ args.datasets_eval = [
53
+ 'KITTI',
54
+ 'NYUv2',
55
+ 'ARKitScenes',
56
+ 'Waymo',
57
+ 'RGBD',
58
+ 'ScanNet',
59
+ 'MVS',
60
+ ]
61
+
62
+ args.load_ckpt = os.path.join(project_root, 'model_zoo/Release', 'wild_camera_all.pth')
63
+ main_worker(args)
external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_megadepth.py ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, sys, inspect, pickle
2
+ project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
3
+ sys.path.insert(0, project_root)
4
+
5
+ import numpy as np
6
+ import torch
7
+ from tabulate import tabulate
8
+ from PIL import Image
9
+ from tqdm import tqdm
10
+ from argparse import ArgumentParser
11
+ from WildCamera.evaluation.evaluate_pose import compute_pose_error, pose_auc, estimate_pose, compute_relative_pose
12
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
13
+
14
+ class Megadepth1500Benchmark:
15
+ def __init__(self, data_root="data/megadepth", scene_names=None) -> None:
16
+ if scene_names is None:
17
+ self.scene_names = [
18
+ "0015_0.1_0.3.npz",
19
+ "0015_0.3_0.5.npz",
20
+ "0022_0.1_0.3.npz",
21
+ "0022_0.3_0.5.npz",
22
+ "0022_0.5_0.7.npz",
23
+ ]
24
+ else:
25
+ self.scene_names = scene_names
26
+ self.scenes = [
27
+ np.load("{}/scene_info/{}".format(data_root, scene), allow_pickle=True)
28
+ for scene in self.scene_names
29
+ ]
30
+ self.data_root = data_root
31
+
32
+ @torch.no_grad()
33
+ def benchmark(self, camera_calibrator, use_calibrated_intrinsic=False):
34
+ with torch.no_grad():
35
+ data_root = self.data_root
36
+ tot_e_t, tot_e_R, tot_e_pose = [], [], []
37
+ for scene_ind in range(len(self.scenes)):
38
+ scene = self.scenes[scene_ind]
39
+ pairs = scene["pair_infos"]
40
+ intrinsics = scene["intrinsics"]
41
+ poses = scene["poses"]
42
+ im_paths = scene["image_paths"]
43
+ pair_inds = range(len(pairs))
44
+ for pairind in tqdm(pair_inds, disable=False):
45
+ idx1, idx2 = pairs[pairind][0]
46
+ K1 = intrinsics[idx1].copy()
47
+ T1 = poses[idx1].copy()
48
+ R1, t1 = T1[:3, :3], T1[:3, 3]
49
+ K2 = intrinsics[idx2].copy()
50
+ T2 = poses[idx2].copy()
51
+ R2, t2 = T2[:3, :3], T2[:3, 3]
52
+ R, t = compute_relative_pose(R1, t1, R2, t2)
53
+ im1_path = f"{data_root}/{im_paths[idx1]}"
54
+ im2_path = f"{data_root}/{im_paths[idx2]}"
55
+ im1 = Image.open(im1_path)
56
+ w1, h1 = im1.size
57
+ im2 = Image.open(im2_path)
58
+ w2, h2 = im2.size
59
+
60
+ if not use_calibrated_intrinsic:
61
+ K1 = K1.copy()
62
+ K2 = K2.copy()
63
+ else:
64
+ K1, _ = camera_calibrator.inference(im1, wtassumption=True)
65
+ K1 = K1.astype(np.float64)
66
+ K2, _ = camera_calibrator.inference(im2, wtassumption=True)
67
+ K2 = K2.astype(np.float64)
68
+
69
+ scale1 = 1200 / max(w1, h1)
70
+ scale2 = 1200 / max(w2, h2)
71
+ w1, h1 = scale1 * w1, scale1 * h1
72
+ w2, h2 = scale2 * w2, scale2 * h2
73
+ K1[:2] = K1[:2] * scale1
74
+ K2[:2] = K2[:2] * scale2
75
+
76
+ corres_fold = os.path.join(self.data_root, 'MegaDepthCorrespondence')
77
+ sv_path = os.path.join(corres_fold, '{}_{}.pkl'.format(str(scene_ind), str(pairind)))
78
+ with open(sv_path, 'rb') as f:
79
+ sparse_matches = pickle.load(f)
80
+
81
+ kpts1 = sparse_matches[:, :2]
82
+ kpts1 = (
83
+ np.stack(
84
+ (
85
+ w1 * (kpts1[:, 0] + 1) / 2,
86
+ h1 * (kpts1[:, 1] + 1) / 2,
87
+ ),
88
+ axis=-1,
89
+ )
90
+ )
91
+ kpts2 = sparse_matches[:, 2:]
92
+ kpts2 = (
93
+ np.stack(
94
+ (
95
+ w2 * (kpts2[:, 0] + 1) / 2,
96
+ h2 * (kpts2[:, 1] + 1) / 2,
97
+ ),
98
+ axis=-1,
99
+ )
100
+ )
101
+
102
+ for _ in range(5):
103
+ shuffling = np.random.permutation(np.arange(len(kpts1)))
104
+ kpts1 = kpts1[shuffling]
105
+ kpts2 = kpts2[shuffling]
106
+ try:
107
+ norm_threshold = 0.5 / (
108
+ np.mean(np.abs(K1[:2, :2])) + np.mean(np.abs(K2[:2, :2])))
109
+ R_est, t_est, mask = estimate_pose(
110
+ kpts1,
111
+ kpts2,
112
+ K1,
113
+ K2,
114
+ norm_threshold,
115
+ conf=0.99999
116
+ )
117
+ T1_to_2_est = np.concatenate((R_est, t_est), axis=-1) #
118
+ e_t, e_R = compute_pose_error(T1_to_2_est, R, t)
119
+ e_pose = max(e_t, e_R)
120
+ except Exception as e:
121
+ e_t, e_R = 90, 90
122
+ e_pose = max(e_t, e_R)
123
+
124
+ tot_e_t.append(e_t)
125
+ tot_e_R.append(e_R)
126
+ tot_e_pose.append(e_pose)
127
+
128
+ tot_e_pose = np.array(tot_e_pose)
129
+ thresholds = [5, 10, 20]
130
+ auc = pose_auc(tot_e_pose, thresholds)
131
+ acc_5 = (tot_e_pose < 5).mean()
132
+ acc_10 = (tot_e_pose < 10).mean()
133
+ acc_15 = (tot_e_pose < 15).mean()
134
+ acc_20 = (tot_e_pose < 20).mean()
135
+ map_5 = acc_5
136
+ map_10 = np.mean([acc_5, acc_10])
137
+ map_20 = np.mean([acc_5, acc_10, acc_15, acc_20])
138
+
139
+ result = {
140
+ "auc_5": auc[0],
141
+ "auc_10": auc[1],
142
+ "auc_20": auc[2],
143
+ "map_5": map_5,
144
+ "map_10": map_10,
145
+ "map_20": map_20,
146
+ }
147
+
148
+ for k in result.keys():
149
+ result[k] = result[k] * 100
150
+
151
+ return result
152
+
153
+ if __name__ == "__main__":
154
+ parser = ArgumentParser()
155
+ parser.add_argument("--data_path", type=str, required=True)
156
+ parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
157
+ args, _ = parser.parse_known_args()
158
+
159
+ megaloftr_benchmark = Megadepth1500Benchmark(args.data_path)
160
+
161
+ args.load_ckpt = os.path.join(project_root, 'model_zoo', 'Release', 'wild_camera_all.pth')
162
+ camera_calibrator = NEWCRFIF(version='large07', pretrained=None)
163
+ camera_calibrator.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
164
+ camera_calibrator.eval()
165
+ camera_calibrator.cuda()
166
+
167
+ result = megaloftr_benchmark.benchmark(camera_calibrator, use_calibrated_intrinsic=True)
168
+ print(tabulate(result.items(), headers=['Metric', 'Scores'], tablefmt='fancy_grid', floatfmt=".2f", numalign="left"))
external/WildCamera/WildCamera/benchmark/benchmark_uncalibtwoview_scannet.py ADDED
@@ -0,0 +1,165 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, sys, inspect, tqdm, copy, pickle
2
+ project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
3
+ sys.path.insert(0, project_root)
4
+
5
+ import numpy as np
6
+ import torch
7
+ from tabulate import tabulate
8
+ from PIL import Image
9
+ from tqdm import tqdm
10
+ from argparse import ArgumentParser
11
+ from WildCamera.evaluation.evaluate_pose import compute_pose_error, pose_auc, estimate_pose, compute_relative_pose
12
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
13
+
14
+ class ScanNetBenchmark:
15
+ def __init__(self, data_root="data/scannet") -> None:
16
+ self.data_root = data_root
17
+
18
+ def benchmark(self, camera_calibrator, use_calibrated_intrinsic=False):
19
+ with torch.no_grad():
20
+ data_root = self.data_root
21
+ tmp = np.load(os.path.join(data_root, "test.npz"))
22
+ pairs, rel_pose = tmp["name"], tmp["rel_pose"]
23
+ tot_e_t, tot_e_R, tot_e_pose = [], [], []
24
+ np.random.seed(0)
25
+ pair_inds = np.random.choice(
26
+ range(len(pairs)), size=len(pairs), replace=False
27
+ )
28
+ for cnt, pairind in enumerate(tqdm(pair_inds)):
29
+ scene = pairs[pairind]
30
+ scene_name = f"scene0{scene[0]}_00"
31
+ im1_path = os.path.join(
32
+ self.data_root,
33
+ "scans_test",
34
+ scene_name,
35
+ "color",
36
+ f"{scene[2]}.jpg",
37
+ )
38
+ im1 = Image.open(im1_path)
39
+ im2_path = os.path.join(
40
+ self.data_root,
41
+ "scans_test",
42
+ scene_name,
43
+ "color",
44
+ f"{scene[3]}.jpg",
45
+ )
46
+ im2 = Image.open(im2_path)
47
+ T_gt = rel_pose[pairind].reshape(3, 4)
48
+ R, t = T_gt[:3, :3], T_gt[:3, 3]
49
+ K = np.stack(
50
+ [
51
+ np.array([float(i) for i in r.split()])
52
+ for r in open(
53
+ os.path.join(
54
+ self.data_root,
55
+ "scans_test",
56
+ scene_name,
57
+ "intrinsic",
58
+ "intrinsic_color.txt",
59
+ ),
60
+ "r",
61
+ )
62
+ .read()
63
+ .split("\n")
64
+ if r
65
+ ]
66
+ )
67
+ w1, h1 = im1.size
68
+ w2, h2 = im2.size
69
+
70
+ if not use_calibrated_intrinsic:
71
+ K1 = K.copy()
72
+ K2 = K.copy()
73
+ else:
74
+ K1_, _ = camera_calibrator.inference(im1, wtassumption=True)
75
+ K1_ = K1_.astype(np.float64)
76
+ K2_, _ = camera_calibrator.inference(im2, wtassumption=True)
77
+ K2_ = K2_.astype(np.float64)
78
+
79
+ K1 = np.eye(4)
80
+ K1[0:3, 0:3] = K1_
81
+
82
+ K2 = np.eye(4)
83
+ K2[0:3, 0:3] = K2_
84
+
85
+ scale1 = 480 / min(w1, h1)
86
+ scale2 = 480 / min(w2, h2)
87
+ w1, h1 = scale1 * w1, scale1 * h1
88
+ w2, h2 = scale2 * w2, scale2 * h2
89
+ K1 = K1 * scale1
90
+ K2 = K2 * scale2
91
+
92
+ corres_fold = os.path.join(self.data_root, 'ScanNetCorrespondence')
93
+ sv_path = os.path.join(corres_fold, '{}.pkl'.format(str(pairind)))
94
+ with open(sv_path, 'rb') as f:
95
+ sparse_matches = pickle.load(f)
96
+
97
+ kpts1 = sparse_matches[:, :2]
98
+ kpts2 = sparse_matches[:, 2:]
99
+
100
+ for _ in range(5):
101
+ shuffling = np.random.permutation(np.arange(len(kpts1)))
102
+ kpts1 = kpts1[shuffling]
103
+ kpts2 = kpts2[shuffling]
104
+ try:
105
+ norm_threshold = 0.8 / (
106
+ np.mean(np.abs(K1[:2, :2])) + np.mean(np.abs(K2[:2, :2])))
107
+ R_est, t_est, mask = estimate_pose(
108
+ kpts1,
109
+ kpts2,
110
+ K1,
111
+ K2,
112
+ norm_threshold,
113
+ conf=0.99999,
114
+ )
115
+ T1_to_2_est = np.concatenate((R_est, t_est), axis=-1) #
116
+ e_t, e_R = compute_pose_error(T1_to_2_est, R, t)
117
+ e_pose = max(e_t, e_R)
118
+ except Exception as _:
119
+ e_t, e_R = 90, 90
120
+ e_pose = max(e_t, e_R)
121
+ tot_e_t.append(e_t)
122
+ tot_e_R.append(e_R)
123
+ tot_e_pose.append(e_pose)
124
+
125
+ tot_e_pose = np.array(tot_e_pose)
126
+ thresholds = [5, 10, 20]
127
+ auc = pose_auc(tot_e_pose, thresholds)
128
+ acc_5 = (tot_e_pose < 5).mean()
129
+ acc_10 = (tot_e_pose < 10).mean()
130
+ acc_15 = (tot_e_pose < 15).mean()
131
+ acc_20 = (tot_e_pose < 20).mean()
132
+ map_5 = acc_5
133
+ map_10 = np.mean([acc_5, acc_10])
134
+ map_20 = np.mean([acc_5, acc_10, acc_15, acc_20])
135
+
136
+ result = {
137
+ "auc_5": auc[0],
138
+ "auc_10": auc[1],
139
+ "auc_20": auc[2],
140
+ "map_5": map_5,
141
+ "map_10": map_10,
142
+ "map_20": map_20,
143
+ }
144
+
145
+ for k in result.keys():
146
+ result[k] = result[k] * 100
147
+
148
+ return result
149
+
150
+ if __name__ == "__main__":
151
+ parser = ArgumentParser()
152
+ parser.add_argument("--data_path", type=str, required=True)
153
+ parser.add_argument('--load_ckpt', type=str, help='path of ckpt')
154
+ args, _ = parser.parse_known_args()
155
+
156
+ scannet_benchmark = ScanNetBenchmark(args.data_path)
157
+
158
+ args.load_ckpt = os.path.join(project_root, 'model_zoo', 'Release', 'wild_camera_all.pth')
159
+ camera_calibrator = NEWCRFIF(version='large07', pretrained=None)
160
+ camera_calibrator.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
161
+ camera_calibrator.eval()
162
+ camera_calibrator.cuda()
163
+
164
+ result = scannet_benchmark.benchmark(camera_calibrator, use_calibrated_intrinsic=True)
165
+ print(tabulate(result.items(), headers=['Metric', 'Scores'], tablefmt='fancy_grid', floatfmt=".2f", numalign="left"))
external/WildCamera/WildCamera/datasets/GSV.py ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, io, glob, natsort, random, copy
2
+ import h5py
3
+ import torch
4
+ import numpy as np
5
+ from PIL import Image
6
+ from torchvision import transforms
7
+
8
+ class GSV:
9
+ def __init__(self, data_root, ht=384, wt=512, shuffleseed=None, split='train') -> None:
10
+ split_path = os.path.join('splits', 'gsv_{}.txt'.format(split))
11
+
12
+ with open(split_path) as file:
13
+ data_names = [line.rstrip() for line in file]
14
+
15
+ if split == 'train':
16
+ if shuffleseed is not None:
17
+ random.seed(shuffleseed)
18
+ data_names = copy.deepcopy(data_names) + copy.deepcopy(data_names)
19
+ random.shuffle(data_names)
20
+
21
+ self.data_root = data_root
22
+ self.wt, self.ht = wt, ht
23
+ self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
24
+ self.tensor = transforms.ToTensor()
25
+ self.data_names = data_names
26
+
27
+ def __len__(self):
28
+ return len(self.data_names)
29
+
30
+ def load_im(self, im_ref):
31
+ im = Image.open(im_ref)
32
+ return im
33
+
34
+ def __getitem__(self, idx):
35
+ # read intrinsics of original size
36
+ scene_name, entry_name = self.data_names[idx].split(' ')
37
+ h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
38
+
39
+ with h5py.File(h5pypath, 'r') as hf:
40
+ K_color = np.array(hf['intrinsic'][entry_name])
41
+
42
+ # Load positive pair data
43
+ rgb = self.load_im(io.BytesIO(np.array(hf['color'][entry_name])))
44
+ w, h = rgb.size
45
+ rgb = rgb.resize((self.wt, self.ht))
46
+
47
+ scaleM = np.eye(3)
48
+ scaleM[0, 0] = self.wt / w
49
+ scaleM[1, 1] = self.ht / h
50
+ aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
51
+
52
+ K = torch.from_numpy(scaleM @ K_color).float()
53
+
54
+ fovy = 2 * np.arctan(h / K_color[1, 1] / 2)
55
+ fovy = fovy.item()
56
+
57
+ # Recompute camera intrinsic matrix due to the resize
58
+ rgb = self.normalize(self.tensor(rgb))
59
+
60
+ data_dict = {
61
+ 'K': K,
62
+ 'rgb': rgb,
63
+ 'K_raw': torch.clone(K),
64
+ 'rgb_raw': torch.clone(rgb),
65
+ 'aspect_ratio_restoration': aspect_ratio_restoration,
66
+ 'fovy': fovy,
67
+ # Only Required in Crop Evaluation and Visualization in Training
68
+ 'T': torch.eye(3, dtype=torch.float32),
69
+ 'scaleM': scaleM.astype(np.float32),
70
+ 'size_wo_change': (h, w),
71
+ }
72
+
73
+ return data_dict
external/WildCamera/WildCamera/datasets/GenericDataset.py ADDED
@@ -0,0 +1,171 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, io, glob, natsort, random, copy
2
+ import h5py
3
+ import torch
4
+ import numpy as np
5
+ import hashlib
6
+ from PIL import Image
7
+ from torchvision import transforms
8
+ from tools.tools import coords_gridN, resample_rgb, apply_augmentation
9
+ from torchvision.transforms import ColorJitter
10
+
11
+ def add_white_noise(rgb):
12
+ w, h = rgb.size
13
+ rgb = np.array(rgb).astype(np.float32)
14
+ rgb = rgb + np.random.randint(9, size=(h, w, 3)) - 2
15
+ rgb = np.clip(rgb, a_min=0, a_max=255)
16
+ rgb = np.round(rgb).astype(np.uint8)
17
+ rgb = Image.fromarray(rgb)
18
+ return rgb
19
+
20
+ class GenericDataset:
21
+ def __init__(self,
22
+ data_root,
23
+ ht=384, wt=512,
24
+ augmentation=False,
25
+ shuffleseed=None,
26
+ split='train',
27
+ datasetname='MegaDepth',
28
+ augscale=2.0,
29
+ no_change_prob=0.1,
30
+ coloraugmentation=False,
31
+ coloraugmentation_scale=0.1,
32
+ transformcategory='transform_calibration'
33
+ ) -> None:
34
+
35
+ name_mapping = {
36
+ 'ScanNet': 'scannet',
37
+ 'MegaDepth': 'megadepth',
38
+ 'NYUv2': 'nyuv2',
39
+ 'Cityscapes': 'cityscapes',
40
+ 'MVS': 'mvs',
41
+ 'RGBD': 'rgbd',
42
+ 'Scenes11': 'scenes11',
43
+ 'SUN3D': 'sun3d',
44
+ 'BIWIRGBDID': 'biwirgbdid',
45
+ 'CAD120': 'cad120',
46
+ 'KITTI': 'kitti',
47
+ 'Waymo': 'waymo',
48
+ 'Nuscenes': 'nuscenes',
49
+ 'ARKitScenes': 'arkitscenes',
50
+ 'Objectron': 'objectron',
51
+ 'MVImgNet': 'mvimgnet'
52
+ }
53
+ split_path = os.path.join('splits', '{}_{}.txt'.format(name_mapping[datasetname], split))
54
+ with open(split_path) as file:
55
+ data_names = [line.rstrip() for line in file]
56
+
57
+ if split == 'train':
58
+ if shuffleseed is not None:
59
+ random.seed(shuffleseed)
60
+ random.shuffle(data_names)
61
+ self.training = True
62
+ else:
63
+ self.training = False
64
+
65
+ self.data_root = data_root
66
+
67
+ self.wt, self.ht = wt, ht
68
+ self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
69
+ self.tensor = transforms.ToTensor()
70
+
71
+ self.data_names = data_names
72
+ self.augmentation = augmentation
73
+ self.augscale = augscale
74
+ self.no_change_prob = no_change_prob
75
+ self.coloraugmentation = coloraugmentation
76
+ self.coloraugmentation_scale = coloraugmentation_scale
77
+
78
+ self.datasetname = datasetname
79
+ self.transformcategory = transformcategory
80
+
81
+ def __len__(self):
82
+ return len(self.data_names)
83
+
84
+ def load_im(self, im_ref):
85
+ im = Image.open(im_ref)
86
+ return im
87
+
88
+ def color_augmentation_fun(self, rgb):
89
+ if random.uniform(0, 1) > 0.5:
90
+ colorjitter = ColorJitter(
91
+ brightness=self.coloraugmentation_scale,
92
+ contrast=self.coloraugmentation_scale,
93
+ saturation=self.coloraugmentation_scale,
94
+ hue=self.coloraugmentation_scale / 3.14
95
+ )
96
+ rgb = colorjitter(rgb)
97
+ rgb = add_white_noise(rgb)
98
+ return rgb
99
+
100
+ def __getitem__(self, idx):
101
+ # While augmenting novel intrinsic, we follow:
102
+ # Step 1 : Resize from original resolution to 480 x 640 (input height x input width)
103
+ # Step 2 : Apply random spatial augmentation
104
+
105
+ scene_name, stem_name = self.data_names[idx].split(' ')
106
+ h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
107
+
108
+ if not os.path.exists(h5pypath):
109
+ print("H5 file %s missing" % h5pypath)
110
+ assert os.path.exists(h5pypath)
111
+
112
+ with h5py.File(h5pypath, 'r') as hf:
113
+ # Load Intrinsic
114
+ K_color = np.array(hf['intrinsic'][stem_name])
115
+
116
+ # Load RGB
117
+ rgb = self.load_im(io.BytesIO(np.array(hf['color'][stem_name])))
118
+
119
+ w, h = rgb.size
120
+
121
+ # Step 1 : Resize
122
+ rgb = rgb.resize((self.wt, self.ht))
123
+
124
+ scaleM = np.eye(3)
125
+ scaleM[0, 0] = self.wt / w
126
+ scaleM[1, 1] = self.ht / h
127
+ aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
128
+
129
+ K = torch.from_numpy(scaleM @ K_color).float()
130
+
131
+ # Color Augmentation only in training
132
+ rgb = self.color_augmentation_fun(rgb) if self.coloraugmentation else rgb
133
+
134
+ # Normalization
135
+ rgb = self.normalize(self.tensor(rgb))
136
+
137
+ # Save RAW
138
+ K_raw = torch.clone(K)
139
+ rgb_raw = torch.clone(rgb)
140
+
141
+ # Step 2 : Random spatial augmentation
142
+ if self.augmentation:
143
+ if self.training:
144
+ rgb, K, T = apply_augmentation(
145
+ rgb, K, seed=None, augscale=self.augscale, no_change_prob=self.no_change_prob
146
+ )
147
+ else:
148
+ T = np.array(h5py.File(h5pypath, 'r')[self.transformcategory][stem_name])
149
+ T = torch.from_numpy(T).float()
150
+ K = torch.inverse(T) @ K
151
+
152
+ _, h_, w_ = rgb.shape
153
+ rgb = resample_rgb(rgb.unsqueeze(0), T, 1, h_, w_, rgb.device).squeeze(0)
154
+ else:
155
+ T = torch.eye(3, dtype=torch.float32)
156
+
157
+ # Exportation
158
+ data_dict = {
159
+ 'K': K,
160
+ 'rgb': rgb,
161
+ 'K_raw': K_raw,
162
+ 'rgb_raw': rgb_raw,
163
+ 'aspect_ratio_restoration': aspect_ratio_restoration,
164
+ 'datasetname': self.datasetname,
165
+ # Only Required in Crop Evaluation
166
+ 'T': T,
167
+ 'scaleM': scaleM.astype(np.float32),
168
+ 'size_wo_change': (h, w),
169
+ }
170
+
171
+ return data_dict
external/WildCamera/WildCamera/datasets/IncdDataset.py ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import torch
3
+ import numpy as np
4
+
5
+ from WildCamera.datasets.MegaDepth import MegaDepth
6
+ from WildCamera.datasets.GSV import GSV
7
+ from WildCamera.datasets.GenericDataset import GenericDataset
8
+
9
+ from torch.utils.data import Subset
10
+
11
+ def IncdDataset(
12
+ data_root,
13
+ datasets_included,
14
+ ht=384, wt=512,
15
+ augmentation=False,
16
+ shuffleseed=None,
17
+ split='train',
18
+ dataset_favour_long=0.0,
19
+ augscale=2.0,
20
+ no_change_prob=0.1,
21
+ coloraugmentation=False,
22
+ coloraugmentation_scale=0.0,
23
+ transformcategory='transform_crop' # For Evaluaiton Purpose
24
+ ) -> None:
25
+ datasets = dict()
26
+
27
+ for datasetname in datasets_included:
28
+ if datasetname == 'GSV':
29
+ datasets[datasetname] = GSV(
30
+ os.path.join(data_root, 'GSV'),
31
+ ht=ht, wt=wt,
32
+ shuffleseed=shuffleseed,
33
+ split=split
34
+ )
35
+ elif datasetname == 'MegaDepth':
36
+ datasets[datasetname] = MegaDepth(
37
+ os.path.join(data_root, 'MegaDepth'),
38
+ ht=ht, wt=wt,
39
+ shuffleseed=shuffleseed,
40
+ split=split,
41
+ )
42
+ elif datasetname == 'BIWIRGBDID':
43
+ datasets[datasetname] = GenericDataset(
44
+ os.path.join(data_root, 'BIWIRGBDID'),
45
+ ht=ht, wt=wt,
46
+ augmentation=False,
47
+ shuffleseed=shuffleseed,
48
+ split=split,
49
+ datasetname=datasetname,
50
+ coloraugmentation=coloraugmentation,
51
+ coloraugmentation_scale=coloraugmentation_scale
52
+ )
53
+ elif datasetname == 'CAD120':
54
+ datasets[datasetname] = GenericDataset(
55
+ os.path.join(data_root, 'CAD120'),
56
+ ht=ht, wt=wt,
57
+ augmentation=False,
58
+ shuffleseed=shuffleseed,
59
+ split=split,
60
+ datasetname=datasetname,
61
+ coloraugmentation=coloraugmentation,
62
+ coloraugmentation_scale=coloraugmentation_scale
63
+ )
64
+ elif (datasetname == 'Objectron') or (datasetname == 'MVImgNet'):
65
+ augscale_obj = 1 + (augscale - 1) / 5
66
+ datasets[datasetname] = GenericDataset(
67
+ os.path.join(data_root, datasetname),
68
+ ht=ht, wt=wt,
69
+ augmentation=True,
70
+ shuffleseed=shuffleseed,
71
+ split=split,
72
+ datasetname=datasetname,
73
+ augscale=augscale_obj,
74
+ no_change_prob=no_change_prob,
75
+ coloraugmentation=coloraugmentation,
76
+ coloraugmentation_scale=coloraugmentation_scale
77
+ )
78
+ else:
79
+ datasets[datasetname] = GenericDataset(
80
+ os.path.join(data_root, datasetname),
81
+ ht=ht, wt=wt,
82
+ augmentation=augmentation,
83
+ shuffleseed=shuffleseed,
84
+ split=split,
85
+ datasetname=datasetname,
86
+ augscale=augscale,
87
+ no_change_prob=no_change_prob,
88
+ coloraugmentation=coloraugmentation,
89
+ coloraugmentation_scale=coloraugmentation_scale,
90
+ transformcategory=transformcategory
91
+ )
92
+
93
+ if split == 'train':
94
+ min_len = np.array([len(datasets[key]) for key in datasets.keys()]).min()
95
+
96
+ datasets_sub = dict()
97
+ for key in datasets.keys():
98
+ dataset = datasets[key]
99
+ subsample_len = np.floor((1 - dataset_favour_long) * min_len + dataset_favour_long * len(dataset))
100
+ subsample_len = int(subsample_len)
101
+
102
+ indices = torch.randperm(len(dataset))
103
+ subsampled_dataset = Subset(dataset, indices[0:subsample_len])
104
+ datasets_sub[key] = subsampled_dataset
105
+
106
+ incdataset = torch.utils.data.ConcatDataset([datasets_sub[key] for key in datasets_sub.keys()])
107
+ else:
108
+ incdataset = torch.utils.data.ConcatDataset([datasets[key] for key in datasets.keys()])
109
+
110
+ return incdataset
external/WildCamera/WildCamera/datasets/MegaDepth.py ADDED
@@ -0,0 +1,76 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, io, glob, natsort, random
2
+ import h5py
3
+ import torch
4
+ import numpy as np
5
+ from PIL import Image
6
+ from torchvision import transforms
7
+
8
+ class MegaDepth:
9
+ def __init__(self, data_root, ht=384, wt=512, shuffleseed=None, split='train') -> None:
10
+ split_path = os.path.join('splits', 'megadepth_{}.txt'.format(split))
11
+
12
+ with open(split_path) as file:
13
+ data_names = [line.rstrip() for line in file]
14
+
15
+ if split == 'train':
16
+ if shuffleseed is not None:
17
+ random.seed(shuffleseed)
18
+ random.shuffle(data_names)
19
+
20
+ self.data_root = data_root
21
+
22
+ self.wt, self.ht = wt, ht
23
+ self.normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
24
+ self.tensor = transforms.ToTensor()
25
+ self.data_names = data_names
26
+
27
+ self.datasetname = 'MegaDepth'
28
+
29
+ def __len__(self):
30
+ return len(self.data_names)
31
+
32
+ def load_im(self, im_ref):
33
+ im = Image.open(im_ref)
34
+ return im
35
+
36
+ def __getitem__(self, idx):
37
+ # read intrinsics of original size
38
+ scene_name, stem_name_color, stem_name_intrinsic = self.data_names[idx].split(' ')
39
+ h5pypath = os.path.join(self.data_root, '{}.hdf5'.format(scene_name))
40
+
41
+ with h5py.File(h5pypath, 'r') as hf:
42
+ K_color = np.array(hf['intrinsic'][stem_name_intrinsic])
43
+
44
+ # Load positive pair data
45
+ rgb = self.load_im(io.BytesIO(np.array(hf['color'][stem_name_color])))
46
+ w, h = rgb.size
47
+ rgb = rgb.resize((self.wt, self.ht))
48
+
49
+ scaleM = np.eye(3)
50
+ scaleM[0, 0] = self.wt / w
51
+ scaleM[1, 1] = self.ht / h
52
+ aspect_ratio_restoration = (scaleM[1, 1] / scaleM[0, 0]).item()
53
+
54
+ K = torch.from_numpy(scaleM @ K_color).float()
55
+
56
+ # Recompute camera intrinsic matrix due to the resize
57
+ rgb = self.normalize(self.tensor(rgb))
58
+
59
+ # Save RAW
60
+ K_raw = torch.clone(K)
61
+ rgb_raw = torch.clone(rgb)
62
+
63
+ data_dict = {
64
+ 'K': K,
65
+ 'rgb': rgb,
66
+ 'K_raw': K_raw,
67
+ 'rgb_raw': rgb_raw,
68
+ 'aspect_ratio_restoration': aspect_ratio_restoration,
69
+ 'datasetname': self.datasetname,
70
+ # Only Required in Crop Evaluation
71
+ 'T': torch.eye(3, dtype=torch.float32),
72
+ 'scaleM': scaleM.astype(np.float32),
73
+ 'size_wo_change': (h, w),
74
+ }
75
+
76
+ return data_dict
external/WildCamera/WildCamera/evaluation/evaluate_crop.py ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import sys, os, time, inspect, tabulate
2
+ import tqdm
3
+ import numpy as np
4
+
5
+ import torch
6
+ import torch.distributed as dist
7
+ from torch.utils.data import DataLoader
8
+ from torchvision import transforms, ops
9
+
10
+ from loguru import logger
11
+
12
+ from tools.tools import to_cuda, DistributedSamplerNoEvenlyDivisible
13
+ from tools.visualization import tensor2rgb
14
+ from tools.calibrator import MonocularCalibrator
15
+ from WildCamera.datasets.IncdDataset import IncdDataset
16
+
17
+ class EvaluateCrop:
18
+ def __init__(self):
19
+ return
20
+
21
+ def evaluate(self, model, args, group=None):
22
+ measurements_all = dict()
23
+
24
+ for dataaset in args.datasets_eval:
25
+
26
+ augmentation = True
27
+
28
+ val_dataset = IncdDataset(
29
+ data_root=args.data_path,
30
+ datasets_included=[dataaset],
31
+ ht=args.input_height,
32
+ wt=args.input_width,
33
+ augmentation=augmentation,
34
+ shuffleseed=None,
35
+ split='test',
36
+ augscale=args.augscale,
37
+ no_change_prob=0.5,
38
+ transformcategory='transform_crop'
39
+ )
40
+ measurements = evaluate_crop(model, val_dataset, args, group)
41
+ measurements_all[dataaset] = measurements
42
+
43
+ if args.gpu == 0:
44
+ errors_all_mean = [['Datasets', 'miou', 'acc']]
45
+ for key in measurements_all.keys():
46
+ measurements = measurements_all[key]
47
+ errors_all_mean.append(
48
+ [
49
+ key,
50
+ measurements['miou'],
51
+ measurements['acc'],
52
+ ]
53
+ )
54
+ logger.info(
55
+ "\n" +
56
+ tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
57
+
58
+ torch.cuda.synchronize()
59
+ return
60
+
61
+
62
+ def acquire_bbox(K, h, w):
63
+ pts1 = torch.from_numpy(np.array([[0.5, 0.5, 1]])).float().cuda().T
64
+ pts2 = torch.from_numpy(np.array([[w - 0.5, 0.5, 1]])).float().cuda().T
65
+ pts3 = torch.from_numpy(np.array([[w - 0.5, h - 0.5, 1]])).float().cuda().T
66
+ pts4 = torch.from_numpy(np.array([[0.5, h - 0.5, 1]])).float().cuda().T
67
+
68
+ pts1 = K @ pts1
69
+ pts2 = K @ pts2
70
+ pts3 = K @ pts3
71
+ pts4 = K @ pts4
72
+
73
+ ptss = torch.cat([pts1, pts2, pts3, pts4])
74
+ ptss = ptss.squeeze().cpu().numpy()[:, 0:2]
75
+
76
+ ptss_ = [
77
+ min(ptss[:, 0]),
78
+ min(ptss[:, 1]),
79
+ max(ptss[:, 0]),
80
+ max(ptss[:, 1]),
81
+ ]
82
+ ptss_ = np.array(ptss_)
83
+ ptss_ = torch.from_numpy(ptss_).float().cuda()
84
+ return ptss_
85
+
86
+ def bbox_auc(errors, thresholds):
87
+ sort_idx = np.argsort(errors)
88
+ errors = np.array(errors.copy())[sort_idx]
89
+ recall = (np.arange(len(errors)) + 1) / len(errors)
90
+ errors = np.r_[0.0, errors]
91
+ recall = np.r_[0.0, recall]
92
+ aucs = []
93
+ for t in thresholds:
94
+ last_index = np.searchsorted(errors, t)
95
+ r = np.r_[recall[:last_index], recall[last_index - 1]]
96
+ e = np.r_[errors[:last_index], t]
97
+ aucs.append(np.trapz(r, x=e) / t)
98
+ return aucs
99
+
100
+ @torch.no_grad()
101
+ def evaluate_crop(model, val_dataset, args, group=None):
102
+ """ In Validation, random sample N Images to mimic a real test set situation """
103
+ monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
104
+ model.eval()
105
+ if group is not None:
106
+ sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
107
+ val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
108
+ else:
109
+ val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
110
+
111
+ measurements = torch.zeros(len(val_dataset), 2).cuda(device=args.gpu)
112
+ measurements_detection = torch.zeros(len(val_dataset)).cuda(device=args.gpu)
113
+ world_size = args.world_size
114
+
115
+ th = 0.2
116
+
117
+ for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
118
+ sample_batched = to_cuda(data_blob)
119
+ rgb, K, scaleM, T, K_raw, rgb_raw = sample_batched['rgb'], sample_batched['K'], sample_batched['scaleM'], sample_batched['T'], sample_batched['K_raw'], sample_batched['rgb_raw']
120
+
121
+ incidence = model(rgb)
122
+ Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
123
+
124
+ Kest_wo_change = torch.inverse(scaleM) @ Kest
125
+ fx_fy = (Kest_wo_change[0, 0, 0] / Kest_wo_change[0, 1, 1]).item()
126
+ fx_fy = max(fx_fy, 1 / fx_fy)
127
+ fx_fy = abs(fx_fy - 1)
128
+ diffx = abs(Kest_wo_change[0, 0, 2].item() - sample_batched['size_wo_change'][1].item() / 2)
129
+ diffx = diffx / sample_batched['size_wo_change'][1].item()
130
+ diffy = abs(Kest_wo_change[0, 1, 2].item() - sample_batched['size_wo_change'][0].item() / 2)
131
+ diffy = diffy / sample_batched['size_wo_change'][0].item()
132
+ diffb = max(diffx, diffy)
133
+ genuen = (fx_fy < th) and (diffb < th)
134
+
135
+ genuen_gt = ((T[0].cpu() - torch.eye(3)).abs().max() < 1e-2).item()
136
+
137
+ if genuen == genuen_gt:
138
+ measurements_detection[val_id * world_size + args.gpu] = 1
139
+
140
+ b, _, h, w = rgb.shape
141
+ bbox_est = acquire_bbox(K_raw @ torch.inverse(Kest), h, w)
142
+ bbox_gt = acquire_bbox(K_raw @ torch.inverse(Kgt), h, w)
143
+
144
+ iou = ops.box_iou(bbox_gt.unsqueeze(0), bbox_est.unsqueeze(0))
145
+ measurements[val_id * world_size + args.gpu, 0] = iou.squeeze().item()
146
+ measurements[val_id * world_size + args.gpu, 1] = 1 - float(genuen_gt)
147
+
148
+ if group is not None:
149
+ dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
150
+
151
+ measurements = measurements.cpu().numpy()
152
+ miou = np.sum(measurements[:, 0] * measurements[:, 1]) / np.sum(measurements[:, 1])
153
+ measurements = {
154
+ 'miou': miou,
155
+ 'acc': measurements_detection.mean().item()
156
+ }
157
+ return measurements
external/WildCamera/WildCamera/evaluation/evaluate_fov.py ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import sys, os, time, inspect, tabulate
2
+ import tqdm
3
+ import numpy as np
4
+
5
+ import torch
6
+ import torch.distributed as dist
7
+ from torch.utils.data import DataLoader
8
+
9
+ from loguru import logger
10
+
11
+ from tools.tools import to_cuda
12
+ from tools.calibrator import MonocularCalibrator
13
+ from tools.tools import DistributedSamplerNoEvenlyDivisible
14
+ from WildCamera.datasets.IncdDataset import IncdDataset
15
+
16
+ class EvaluateFov:
17
+
18
+ def __init__(self):
19
+ self.min_fovy = np.inf
20
+
21
+ def evaluate(self, model, args, steps, writer=None, group=None, wtassumption=False, split='test'):
22
+
23
+ if len(args.datasets_eval) > 1:
24
+ autosave = False
25
+ else:
26
+ autosave = True
27
+
28
+ measurements_all = dict()
29
+
30
+ for dataaset in args.datasets_eval:
31
+
32
+ if dataaset in args.datasets_train:
33
+ augmentation = True
34
+ else:
35
+ augmentation = False
36
+
37
+ val_dataset = IncdDataset(
38
+ data_root=args.data_path,
39
+ datasets_included=[dataaset],
40
+ ht=args.input_height,
41
+ wt=args.input_width,
42
+ augmentation=augmentation,
43
+ shuffleseed=None,
44
+ split=split,
45
+ augscale=args.augscale,
46
+ no_change_prob=0.0
47
+ )
48
+ measurements = evaluate_fov(model, val_dataset, args, group, wtassumption=wtassumption)
49
+ measurements_all[dataaset] = measurements
50
+
51
+ if args.gpu == 0:
52
+ if writer is not None and (steps > 0):
53
+ for k in measurements.keys():
54
+ writer.add_scalar('Eval_{}/{}'.format(dataaset, k), measurements[k], steps)
55
+
56
+ if args.gpu == 0 and autosave and (steps > 0):
57
+ if measurements['error_fovy'] < self.min_fovy:
58
+ self.min_fovy = measurements['error_fovy']
59
+ svpath = os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'min_fovy.ckpt')
60
+ try:
61
+ model_ = model.module
62
+ except:
63
+ model_ = model
64
+ torch.save(model_.state_dict(), svpath)
65
+ logger.info("Save to %s" % svpath)
66
+
67
+ if args.gpu == 0 and (not autosave) and (steps > 0):
68
+ try:
69
+ model_ = model.module
70
+ except:
71
+ model_ = model
72
+ torch.save(
73
+ model_.state_dict(),
74
+ os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'step_{}.ckpt'.format(str(steps)))
75
+ )
76
+
77
+ if args.gpu == 0:
78
+ errors_all_mean = [['Datasets', 'error_fovy', 'error_fovy_medi']]
79
+ for key in measurements_all.keys():
80
+ measurements = measurements_all[key]
81
+ errors_all_mean.append(
82
+ [
83
+ key,
84
+ measurements['error_fovy'],
85
+ measurements['error_fovy_medi'],
86
+ ]
87
+ )
88
+ logger.info(
89
+ "\n==================== Performance at Step %d with Assumption %d ====================\n" % (steps, wtassumption) +
90
+ tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
91
+
92
+ torch.cuda.synchronize()
93
+ return
94
+
95
+ @torch.no_grad()
96
+ def evaluate_fov(model, val_dataset, args, group=None, wtassumption=False):
97
+ """ In Validation, random sample N Images to mimic a real test set situation """
98
+ monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
99
+ model.eval()
100
+ if args.world_size > 1:
101
+ sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
102
+ val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
103
+ else:
104
+ val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
105
+
106
+ measurements = torch.zeros(len(val_dataset), 6).cuda(device=args.gpu)
107
+ world_size = args.world_size
108
+
109
+ for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
110
+ sample_batched = to_cuda(data_blob)
111
+ rgb, K, r = sample_batched['rgb'], sample_batched['K'], sample_batched['aspect_ratio_restoration']
112
+
113
+ incidence = model(rgb)
114
+ if not wtassumption:
115
+ Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
116
+ else:
117
+ Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_1DoF(incidence, r=r.item())
118
+
119
+ b, _, h, w = rgb.shape
120
+ device = rgb.device
121
+
122
+ fovy_est, fovy_gt = 2 * torch.arctan(h / Kest[1, 1] / 2), sample_batched['fovy']
123
+ error_fovy = np.abs(np.degrees(fovy_est.item()) - np.degrees(fovy_gt.item()))
124
+
125
+ error_all = np.array([error_fovy])
126
+ error_all = torch.from_numpy(error_all).float().to(device)
127
+
128
+ measurements[val_id * world_size + args.gpu] += error_all
129
+
130
+ if args.world_size > 1:
131
+ dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
132
+
133
+ zero_entry = torch.sum(measurements.abs(), dim=1) == 0
134
+ assert torch.sum(zero_entry) == 0
135
+
136
+ fovy_median = torch.median(measurements.squeeze())
137
+ measurements = torch.mean(measurements, dim=0)
138
+ measurements = {
139
+ 'error_fovy': measurements[0].item(),
140
+ 'error_fovy_medi': fovy_median.item()
141
+ }
142
+ return measurements
external/WildCamera/WildCamera/evaluation/evaluate_intrinsic.py ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import sys, os, time, inspect, tabulate
2
+ import tqdm
3
+ import numpy as np
4
+
5
+ import torch
6
+ import torch.distributed as dist
7
+ from torch.utils.data import DataLoader
8
+
9
+ from loguru import logger
10
+
11
+ from tools.tools import to_cuda
12
+ from tools.calibrator import MonocularCalibrator
13
+ from tools.evaluation import compute_intrinsic_measure
14
+ from tools.tools import DistributedSamplerNoEvenlyDivisible
15
+ from WildCamera.datasets.IncdDataset import IncdDataset
16
+
17
+ class EvaluateIntrinsic:
18
+ def __init__(self):
19
+ self.min_focal = 1e10
20
+
21
+ def evaluate(self, model, args, steps, writer=None, group=None, wtassumption=False):
22
+
23
+ if len(args.datasets_eval) > 1:
24
+ autosave = False
25
+ else:
26
+ autosave = True
27
+
28
+ measurements_all = dict()
29
+
30
+ for dataaset in args.datasets_eval:
31
+
32
+ if dataaset in args.datasets_train:
33
+ augmentation = True
34
+ else:
35
+ augmentation = False
36
+
37
+ val_dataset = IncdDataset(
38
+ data_root=args.data_path,
39
+ datasets_included=[dataaset],
40
+ ht=args.input_height,
41
+ wt=args.input_width,
42
+ augmentation=augmentation,
43
+ shuffleseed=None,
44
+ split='test',
45
+ augscale=args.augscale,
46
+ no_change_prob=0.0,
47
+ transformcategory='transform_calibration'
48
+ )
49
+ measurements = evaluate_intrinsic(model, val_dataset, args, group, wtassumption=wtassumption)
50
+ measurements_all[dataaset] = measurements
51
+
52
+ if args.gpu == 0:
53
+ if writer is not None and (steps > 0):
54
+ for k in measurements.keys():
55
+ writer.add_scalar('Eval_{}/{}'.format(dataaset, k), measurements[k], steps)
56
+
57
+ if args.gpu == 0 and autosave and (steps > 0):
58
+ if measurements['error_f'] < self.min_focal:
59
+ self.min_focal = measurements['error_f']
60
+ svpath = os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'min_focal.ckpt')
61
+ torch.save(model.state_dict(), svpath)
62
+ logger.info("Save to %s" % svpath)
63
+
64
+ if args.gpu == 0 and (not autosave) and (steps > 0):
65
+ torch.save(model.state_dict(), os.path.join(args.saving_location, 'model_zoo', args.experiment_name, 'step_{}.ckpt'.format(str(steps))))
66
+
67
+ if args.gpu == 0:
68
+ errors_all_mean = [['Datasets', 'error_fx', 'error_fy', 'error_f', 'error_bx', 'error_by', 'error_b']]
69
+ for key in measurements_all.keys():
70
+ measurements = measurements_all[key]
71
+ errors_all_mean.append(
72
+ [
73
+ key,
74
+ measurements['error_fx'],
75
+ measurements['error_fy'],
76
+ measurements['error_f'],
77
+ measurements['error_bx'],
78
+ measurements['error_by'],
79
+ measurements['error_b']
80
+ ]
81
+ )
82
+ logger.info(
83
+ "\n==================== Performance at Step %d with Assumption %d ====================\n" % (steps, wtassumption) +
84
+ tabulate.tabulate(errors_all_mean, headers='firstrow', tablefmt='fancy_grid', numalign="center", floatfmt=".3f"))
85
+
86
+ torch.cuda.synchronize()
87
+ return
88
+
89
+ @torch.no_grad()
90
+ def evaluate_intrinsic(model, val_dataset, args, group=None, wtassumption=False):
91
+ """ In Validation, random sample N Images to mimic a real test set situation """
92
+ monocalibrator = MonocularCalibrator(l1_th=args.l1_th)
93
+ model.eval()
94
+ if args.world_size > 1:
95
+ sampler = DistributedSamplerNoEvenlyDivisible(val_dataset, shuffle=False)
96
+ val_loader = DataLoader(val_dataset, batch_size=1, sampler=sampler, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
97
+ else:
98
+ val_loader = DataLoader(val_dataset, batch_size=1, pin_memory=False, shuffle=False, num_workers=args.eval_workers, drop_last=False)
99
+
100
+ measurements = torch.zeros(len(val_dataset), 6).cuda(device=args.gpu)
101
+ world_size = args.world_size
102
+
103
+ for val_id, data_blob in enumerate(tqdm.tqdm(val_loader, disable=False)):
104
+ sample_batched = to_cuda(data_blob)
105
+ rgb, K, r = sample_batched['rgb'], sample_batched['K'], sample_batched['aspect_ratio_restoration']
106
+
107
+ incidence = model(rgb)
108
+ if not wtassumption:
109
+ Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_4DoF(incidence, RANSAC_trial=2048)
110
+ else:
111
+ Kgt, Kest = torch.clone(K).squeeze(0), monocalibrator.calibrate_camera_1DoF(incidence, r=r.item())
112
+
113
+ b, _, h, w = rgb.shape
114
+ device = rgb.device
115
+
116
+ error_fx, error_fy, error_f, error_bx, error_by, error_b = compute_intrinsic_measure(
117
+ Kest=torch.clone(Kest),
118
+ Kgt=torch.clone(Kgt),
119
+ h=h,
120
+ w=w
121
+ )
122
+
123
+ error_all = np.array([error_fx, error_fy, error_f, error_bx, error_by, error_b])
124
+ error_all = torch.from_numpy(error_all).float().to(device)
125
+
126
+ measurements[val_id * world_size + args.gpu] += error_all
127
+
128
+ if args.world_size > 1:
129
+ dist.all_reduce(tensor=measurements, op=dist.ReduceOp.SUM, group=group)
130
+
131
+ zero_entry = torch.sum(measurements.abs(), dim=1) == 0
132
+ assert torch.sum(zero_entry) == 0
133
+
134
+ measurements = torch.mean(measurements, dim=0)
135
+ measurements = {
136
+ 'error_fx': measurements[0].item(),
137
+ 'error_fy': measurements[1].item(),
138
+ 'error_f': measurements[2].item(),
139
+ 'error_bx': measurements[3].item(),
140
+ 'error_by': measurements[4].item(),
141
+ 'error_b': measurements[5].item()
142
+ }
143
+
144
+ return measurements
external/WildCamera/WildCamera/evaluation/evaluate_pose.py ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Put most Common Functions Here
2
+ import cv2
3
+ import torch
4
+ import numpy as np
5
+ import torch.nn.functional as F
6
+
7
+ from einops.einops import rearrange
8
+
9
+ # -- # Pose Benchmark
10
+ def angle_error_mat(R1, R2):
11
+ cos = (np.trace(np.dot(R1.T, R2)) - 1) / 2
12
+ cos = np.clip(cos, -1.0, 1.0) # numercial errors can make it out of bounds
13
+ return np.rad2deg(np.abs(np.arccos(cos)))
14
+
15
+ def angle_error_vec(v1, v2):
16
+ n = np.linalg.norm(v1) * np.linalg.norm(v2)
17
+ return np.rad2deg(np.arccos(np.clip(np.dot(v1, v2) / n, -1.0, 1.0)))
18
+
19
+ def compute_pose_error(T_0to1, R, t):
20
+ R_gt = T_0to1[:3, :3]
21
+ t_gt = T_0to1[:3, 3]
22
+ error_t = angle_error_vec(t.squeeze(), t_gt)
23
+ error_t = np.minimum(error_t, 180 - error_t) # ambiguity of E estimation
24
+ error_R = angle_error_mat(R, R_gt)
25
+ return error_t, error_R
26
+
27
+ def pose_auc(errors, thresholds):
28
+ sort_idx = np.argsort(errors)
29
+ errors = np.array(errors.copy())[sort_idx]
30
+ recall = (np.arange(len(errors)) + 1) / len(errors)
31
+ errors = np.r_[0.0, errors]
32
+ recall = np.r_[0.0, recall]
33
+ aucs = []
34
+ for t in thresholds:
35
+ last_index = np.searchsorted(errors, t)
36
+ r = np.r_[recall[:last_index], recall[last_index - 1]]
37
+ e = np.r_[errors[:last_index], t]
38
+ aucs.append(np.trapz(r, x=e) / t)
39
+ return aucs
40
+
41
+ def compute_relative_pose(R1, t1, R2, t2):
42
+ rots = R2 @ (R1.T)
43
+ trans = -rots @ t1 + t2
44
+ return rots, trans
45
+
46
+ def rotate_intrinsic(K, n):
47
+ base_rot = np.array([[0, 1, 0], [-1, 0, 0], [0, 0, 1]])
48
+ rot = np.linalg.matrix_power(base_rot, n)
49
+ return rot @ K
50
+
51
+ def estimate_pose(kpts0, kpts1, K0, K1, norm_thresh, conf=0.99999):
52
+ if len(kpts0) < 5:
53
+ return None
54
+ K0inv = np.linalg.inv(K0[:2,:2])
55
+ K1inv = np.linalg.inv(K1[:2,:2])
56
+
57
+ kpts0 = (K0inv @ (kpts0-K0[None,:2,2]).T).T
58
+ kpts1 = (K1inv @ (kpts1-K1[None,:2,2]).T).T
59
+
60
+ E, mask = cv2.findEssentialMat(
61
+ kpts0, kpts1, np.eye(3), threshold=norm_thresh, prob=conf, method=cv2.RANSAC
62
+ )
63
+
64
+ ret = None
65
+ if E is not None:
66
+ best_num_inliers = 0
67
+
68
+ for _E in np.split(E, len(E) / 3):
69
+ n, R, t, _ = cv2.recoverPose(_E, kpts0, kpts1, np.eye(3), 1e9, mask=mask)
70
+ if n > best_num_inliers:
71
+ best_num_inliers = n
72
+ ret = (R, t, mask.ravel() > 0)
73
+ return ret
74
+
75
+
76
+ # -- # Compute Error in EPE
77
+
78
+ def compute_flow_metrics(depth1, depth2, T_1to2, K1, K2, dense_matches):
79
+ """[summary]
80
+
81
+ Args:
82
+ H ([type]): [description]
83
+ scale ([type]): [description]
84
+
85
+ Returns:
86
+ [type]: [description]
87
+ """
88
+ dense_matches = rearrange(dense_matches, "b d h w -> b h w d")
89
+ b, h1, w1, d = dense_matches.shape
90
+ x1_n = torch.meshgrid(
91
+ *[
92
+ torch.linspace(
93
+ -1 + 1 / n, 1 - 1 / n, n, device=dense_matches.device
94
+ )
95
+ for n in (b, h1, w1)
96
+ ]
97
+ )
98
+ x1_n = torch.stack((x1_n[2], x1_n[1]), dim=-1).reshape(b, h1 * w1, 2)
99
+ mask, x2 = warp_kpts(
100
+ x1_n.double(),
101
+ depth1.double(),
102
+ depth2.double(),
103
+ T_1to2.double(),
104
+ K1.double(),
105
+ K2.double(),
106
+ )
107
+ mask_validation = mask.float().reshape(b, h1, w1) > 0
108
+ x2 = x2.reshape(b, h1, w1, 2)
109
+
110
+ dense_matches = unnorm_coords_Numpystyle(
111
+ rearrange(dense_matches, "b h w d -> b d h w"),
112
+ h=h1, w=w1
113
+ )
114
+
115
+ x2 = unnorm_coords_Numpystyle(
116
+ rearrange(x2, "b h w d -> b d h w"),
117
+ h=h1, w=w1
118
+ )
119
+
120
+ epe = (dense_matches - x2).norm(dim=1) # *scale?
121
+
122
+ evaluated_pixel_num = torch.sum(mask_validation)
123
+ epe_sum = torch.sum(epe[mask_validation])
124
+ px1 = torch.sum((epe < 1)[mask_validation])
125
+ px5 = torch.sum((epe < 5)[mask_validation])
126
+ px8 = torch.sum((epe < 8)[mask_validation])
127
+
128
+ sub_measurements = torch.stack([epe_sum, px1, px5, px8, evaluated_pixel_num])
129
+ return sub_measurements
130
+
131
+
132
+
133
+ @torch.no_grad()
134
+ def warp_kpts(kpts0, depth0, depth1, T_0to1, K0, K1):
135
+ """Warp kpts0 from I0 to I1 with depth, K and Rt
136
+ Also check covisibility and depth consistency.
137
+ Depth is consistent if relative error < 0.2 (hard-coded).
138
+ # https://github.com/zju3dv/LoFTR/blob/94e98b695be18acb43d5d3250f52226a8e36f839/src/loftr/utils/geometry.py adapted from here
139
+ Args:
140
+ kpts0 (torch.Tensor): [N, L, 2] - <x, y>, should be normalized in (-1,1)
141
+ depth0 (torch.Tensor): [N, H, W],
142
+ depth1 (torch.Tensor): [N, H, W],
143
+ T_0to1 (torch.Tensor): [N, 3, 4],
144
+ K0 (torch.Tensor): [N, 3, 3],
145
+ K1 (torch.Tensor): [N, 3, 3],
146
+ Returns:
147
+ calculable_mask (torch.Tensor): [N, L]
148
+ warped_keypoints0 (torch.Tensor): [N, L, 2] <x0_hat, y1_hat>
149
+ """
150
+ (
151
+ n,
152
+ h,
153
+ w,
154
+ ) = depth0.shape
155
+ kpts0_depth = F.grid_sample(depth0[:, None], kpts0[:, :, None], mode="bilinear", align_corners=False)[
156
+ :, 0, :, 0
157
+ ]
158
+ kpts0 = torch.stack(
159
+ (w * (kpts0[..., 0] + 1) / 2, h * (kpts0[..., 1] + 1) / 2), dim=-1
160
+ ) # [-1+1/h, 1-1/h] -> [0.5, h-0.5]
161
+ # Sample depth, get calculable_mask on depth != 0
162
+ nonzero_mask = kpts0_depth != 0
163
+
164
+ # Unproject
165
+ kpts0_h = (
166
+ torch.cat([kpts0, torch.ones_like(kpts0[:, :, [0]])], dim=-1)
167
+ * kpts0_depth[..., None]
168
+ ) # (N, L, 3)
169
+ kpts0_n = K0.inverse() @ kpts0_h.transpose(2, 1) # (N, 3, L)
170
+ kpts0_cam = kpts0_n
171
+
172
+ # Rigid Transform
173
+ w_kpts0_cam = T_0to1[:, :3, :3] @ kpts0_cam + T_0to1[:, :3, [3]] # (N, 3, L)
174
+ w_kpts0_depth_computed = w_kpts0_cam[:, 2, :]
175
+
176
+ # Project
177
+ w_kpts0_h = (K1 @ w_kpts0_cam).transpose(2, 1) # (N, L, 3)
178
+ w_kpts0 = w_kpts0_h[:, :, :2] / (
179
+ w_kpts0_h[:, :, [2]] + 1e-4
180
+ ) # (N, L, 2), +1e-4 to avoid zero depth
181
+
182
+ # Covisible Check
183
+ h, w = depth1.shape[1:3]
184
+ covisible_mask = (
185
+ (w_kpts0[:, :, 0] > 0)
186
+ * (w_kpts0[:, :, 0] < w - 1)
187
+ * (w_kpts0[:, :, 1] > 0)
188
+ * (w_kpts0[:, :, 1] < h - 1)
189
+ )
190
+ w_kpts0 = torch.stack(
191
+ (2 * w_kpts0[..., 0] / w - 1, 2 * w_kpts0[..., 1] / h - 1), dim=-1
192
+ ) # from [0.5,h-0.5] -> [-1+1/h, 1-1/h]
193
+ # w_kpts0[~covisible_mask, :] = -5 # xd
194
+
195
+ w_kpts0_depth = F.grid_sample(
196
+ depth1[:, None], w_kpts0[:, :, None], mode="bilinear", align_corners=False
197
+ )[:, 0, :, 0]
198
+ consistent_mask = (
199
+ (w_kpts0_depth - w_kpts0_depth_computed) / w_kpts0_depth
200
+ ).abs() < 0.05
201
+ valid_mask = nonzero_mask * covisible_mask * consistent_mask
202
+
203
+ return valid_mask, w_kpts0
204
+
205
+ def unnorm_coords_Numpystyle(coords1, h, w):
206
+ coords1x, coords1y = torch.split(coords1, 1, dim=1)
207
+ coords1x = (coords1x + 1) / 2 * w
208
+ coords1y = (coords1y + 1) / 2 * h
209
+ return torch.cat([coords1x, coords1y], dim=1)
external/WildCamera/WildCamera/newcrfs/__init__.py ADDED
File without changes
external/WildCamera/WildCamera/newcrfs/__pycache__/swin_transformer.cpython-310.pyc ADDED
Binary file (19.4 kB). View file
 
external/WildCamera/WildCamera/newcrfs/__pycache__/uper_crf_head.cpython-310.pyc ADDED
Binary file (11.8 kB). View file
 
external/WildCamera/WildCamera/newcrfs/newcrf_incidencefield.py ADDED
@@ -0,0 +1,196 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+ import torch.nn.functional as F
4
+ import PIL.Image as Image
5
+
6
+ import numpy as np
7
+
8
+ from .swin_transformer import SwinTransformer
9
+ from .newcrf_layers import NewCRF
10
+ from .uper_crf_head import PSP
11
+
12
+ from torchvision import transforms
13
+ from tools.calibrator import MonocularCalibrator
14
+ ########################################################################################################################
15
+
16
+
17
+ class NEWCRFIF(nn.Module):
18
+ """
19
+ Depth network based on neural window FC-CRFs architecture.
20
+ """
21
+ def __init__(self, pretrained=None, frozen_stages=-1, version='large07'):
22
+ super().__init__()
23
+
24
+ self.with_auxiliary_head = False
25
+ self.with_neck = False
26
+
27
+ norm_cfg = dict(type='BN', requires_grad=True)
28
+
29
+ window_size = int(version[-2:])
30
+
31
+ if version[:-2] == 'base':
32
+ embed_dim = 128
33
+ depths = [2, 2, 18, 2]
34
+ num_heads = [4, 8, 16, 32]
35
+ in_channels = [128, 256, 512, 1024]
36
+ elif version[:-2] == 'large':
37
+ embed_dim = 192
38
+ depths = [2, 2, 18, 2]
39
+ num_heads = [6, 12, 24, 48]
40
+ in_channels = [192, 384, 768, 1536]
41
+ elif version[:-2] == 'tiny':
42
+ embed_dim = 96
43
+ depths = [2, 2, 6, 2]
44
+ num_heads = [3, 6, 12, 24]
45
+ in_channels = [96, 192, 384, 768]
46
+
47
+ backbone_cfg = dict(
48
+ embed_dim=embed_dim,
49
+ depths=depths,
50
+ num_heads=num_heads,
51
+ window_size=window_size,
52
+ ape=False,
53
+ drop_path_rate=0.3,
54
+ patch_norm=True,
55
+ use_checkpoint=False,
56
+ frozen_stages=frozen_stages
57
+ )
58
+
59
+ embed_dim = 512
60
+ decoder_cfg = dict(
61
+ in_channels=in_channels,
62
+ in_index=[0, 1, 2, 3],
63
+ pool_scales=(1, 2, 3, 6),
64
+ channels=embed_dim,
65
+ dropout_ratio=0.0,
66
+ num_classes=32,
67
+ norm_cfg=norm_cfg,
68
+ align_corners=False
69
+ )
70
+
71
+ self.backbone = SwinTransformer(**backbone_cfg)
72
+ v_dim = decoder_cfg['num_classes']*4
73
+ win = 7
74
+ crf_dims = [128, 256, 512, 1024]
75
+ v_dims = [64, 128, 256, embed_dim]
76
+ self.crf3 = NewCRF(input_dim=in_channels[3], embed_dim=crf_dims[3], window_size=win, v_dim=v_dims[3], num_heads=32)
77
+ self.crf2 = NewCRF(input_dim=in_channels[2], embed_dim=crf_dims[2], window_size=win, v_dim=v_dims[2], num_heads=16)
78
+ self.crf1 = NewCRF(input_dim=in_channels[1], embed_dim=crf_dims[1], window_size=win, v_dim=v_dims[1], num_heads=8)
79
+ self.crf0 = NewCRF(input_dim=in_channels[0], embed_dim=crf_dims[0], window_size=win, v_dim=v_dims[0], num_heads=4)
80
+
81
+ self.decoder = PSP(**decoder_cfg)
82
+
83
+ self.incidence_head = IncidenceHead(input_dim=crf_dims[0], output_dim=3)
84
+
85
+ self.up_mode = 'bilinear'
86
+ if self.up_mode == 'mask':
87
+ self.mask_head = nn.Sequential(
88
+ nn.Conv2d(crf_dims[0], 64, 3, padding=1),
89
+ nn.ReLU(inplace=True),
90
+ nn.Conv2d(64, 16*9, 1, padding=0))
91
+
92
+ self.init_weights(pretrained=pretrained)
93
+
94
+ def init_weights(self, pretrained=None):
95
+ """Initialize the weights in backbone and heads.
96
+
97
+ Args:
98
+ pretrained (str, optional): Path to pre-trained weights.
99
+ Defaults to None.
100
+ """
101
+ print(f'== Load encoder backbone from: {pretrained}')
102
+ self.backbone.init_weights(pretrained=pretrained)
103
+ self.decoder.init_weights()
104
+ if self.with_auxiliary_head:
105
+ if isinstance(self.auxiliary_head, nn.ModuleList):
106
+ for aux_head in self.auxiliary_head:
107
+ aux_head.init_weights()
108
+ else:
109
+ self.auxiliary_head.init_weights()
110
+
111
+ def upsample_mask(self, disp, mask):
112
+ """ Upsample disp [H/4, W/4, 1] -> [H, W, 1] using convex combination """
113
+ N, _, H, W = disp.shape
114
+ mask = mask.view(N, 1, 9, 4, 4, H, W)
115
+ mask = torch.softmax(mask, dim=2)
116
+
117
+ up_disp = F.unfold(disp, kernel_size=3, padding=1)
118
+ up_disp = up_disp.view(N, 1, 9, 1, 1, H, W)
119
+
120
+ up_disp = torch.sum(mask * up_disp, dim=2)
121
+ up_disp = up_disp.permute(0, 1, 4, 2, 5, 3)
122
+ return up_disp.reshape(N, 1, 4*H, 4*W)
123
+
124
+ def forward(self, imgs):
125
+ feats = self.backbone(imgs)
126
+ if self.with_neck:
127
+ feats = self.neck(feats)
128
+
129
+ ppm_out = self.decoder(feats)
130
+
131
+ e3 = self.crf3(feats[3], ppm_out)
132
+ e3 = nn.PixelShuffle(2)(e3)
133
+ e2 = self.crf2(feats[2], e3)
134
+ e2 = nn.PixelShuffle(2)(e2)
135
+ e1 = self.crf1(feats[1], e2)
136
+ e1 = nn.PixelShuffle(2)(e1)
137
+ e0 = self.crf0(feats[0], e1)
138
+
139
+ incidence = self.incidence_head(e0, 4)
140
+
141
+ return incidence
142
+
143
+ @torch.no_grad()
144
+ def inference(self, rgb, wtassumption=False):
145
+ self.eval()
146
+ input_wt, input_ht = 640, 480
147
+ w, h = rgb.size
148
+
149
+ rgb = rgb.resize((input_wt, input_ht), Image.Resampling.BILINEAR)
150
+
151
+ scaleM = np.eye(3)
152
+ scaleM[0, 0] = input_wt / w
153
+ scaleM[1, 1] = input_ht / h
154
+ scaleM = torch.from_numpy(scaleM).float().to(next(self.parameters()).device)
155
+
156
+ normalize = transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225])
157
+ totensor = transforms.ToTensor()
158
+
159
+ rgb = normalize(totensor(rgb))
160
+ rgb = rgb.unsqueeze(0).to(next(self.parameters()).device)
161
+
162
+ monocalibrator = MonocularCalibrator()
163
+ incidence = self.forward(rgb)
164
+
165
+ if not wtassumption:
166
+ Kest = monocalibrator.calibrate_camera_4DoF(incidence)
167
+ else:
168
+ r = (scaleM[1, 1] / scaleM[0, 0]).item()
169
+ Kest = monocalibrator.calibrate_camera_1DoF(incidence, r=r)
170
+
171
+ Kest = torch.inverse(scaleM) @ Kest
172
+ return Kest.detach().squeeze().cpu().numpy(), incidence
173
+
174
+ @torch.no_grad()
175
+ def restore_image(self, rgb, intrinsic, fixcrop=True):
176
+ monocalibrator = MonocularCalibrator()
177
+ return monocalibrator.restore_image(rgb, intrinsic, fixcrop)
178
+
179
+ class IncidenceHead(nn.Module):
180
+ def __init__(self, input_dim=100, output_dim=3):
181
+ super(IncidenceHead, self).__init__()
182
+ self.conv1 = nn.Conv2d(input_dim, output_dim, 3, padding=1)
183
+
184
+ def forward(self, x, scale):
185
+ x = self.conv1(x)
186
+ if scale > 1:
187
+ x = upsample(x, scale_factor=scale)
188
+
189
+ incidence = torch.nn.functional.normalize(x, dim=1)
190
+ return incidence
191
+
192
+
193
+ def upsample(x, scale_factor=2, mode="bilinear", align_corners=False):
194
+ """Upsample input tensor by a factor of 2
195
+ """
196
+ return F.interpolate(x, scale_factor=scale_factor, mode=mode, align_corners=align_corners)
external/WildCamera/WildCamera/newcrfs/newcrf_layers.py ADDED
@@ -0,0 +1,433 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+ import torch.nn.functional as F
4
+ import torch.utils.checkpoint as checkpoint
5
+ import numpy as np
6
+ from timm.models.layers import DropPath, to_2tuple, trunc_normal_
7
+
8
+
9
+ class Mlp(nn.Module):
10
+ """ Multilayer perceptron."""
11
+
12
+ def __init__(self, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, drop=0.):
13
+ super().__init__()
14
+ out_features = out_features or in_features
15
+ hidden_features = hidden_features or in_features
16
+ self.fc1 = nn.Linear(in_features, hidden_features)
17
+ self.act = act_layer()
18
+ self.fc2 = nn.Linear(hidden_features, out_features)
19
+ self.drop = nn.Dropout(drop)
20
+
21
+ def forward(self, x):
22
+ x = self.fc1(x)
23
+ x = self.act(x)
24
+ x = self.drop(x)
25
+ x = self.fc2(x)
26
+ x = self.drop(x)
27
+ return x
28
+
29
+
30
+ def window_partition(x, window_size):
31
+ """
32
+ Args:
33
+ x: (B, H, W, C)
34
+ window_size (int): window size
35
+
36
+ Returns:
37
+ windows: (num_windows*B, window_size, window_size, C)
38
+ """
39
+ B, H, W, C = x.shape
40
+ x = x.view(B, H // window_size, window_size, W // window_size, window_size, C)
41
+ windows = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(-1, window_size, window_size, C)
42
+ return windows
43
+
44
+
45
+ def window_reverse(windows, window_size, H, W):
46
+ """
47
+ Args:
48
+ windows: (num_windows*B, window_size, window_size, C)
49
+ window_size (int): Window size
50
+ H (int): Height of image
51
+ W (int): Width of image
52
+
53
+ Returns:
54
+ x: (B, H, W, C)
55
+ """
56
+ B = int(windows.shape[0] / (H * W / window_size / window_size))
57
+ x = windows.view(B, H // window_size, W // window_size, window_size, window_size, -1)
58
+ x = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(B, H, W, -1)
59
+ return x
60
+
61
+
62
+ class WindowAttention(nn.Module):
63
+ """ Window based multi-head self attention (W-MSA) module with relative position bias.
64
+ It supports both of shifted and non-shifted window.
65
+
66
+ Args:
67
+ dim (int): Number of input channels.
68
+ window_size (tuple[int]): The height and width of the window.
69
+ num_heads (int): Number of attention heads.
70
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
71
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set
72
+ attn_drop (float, optional): Dropout ratio of attention weight. Default: 0.0
73
+ proj_drop (float, optional): Dropout ratio of output. Default: 0.0
74
+ """
75
+
76
+ def __init__(self, dim, window_size, num_heads, v_dim, qkv_bias=True, qk_scale=None, attn_drop=0., proj_drop=0.):
77
+
78
+ super().__init__()
79
+ self.dim = dim
80
+ self.window_size = window_size # Wh, Ww
81
+ self.num_heads = num_heads
82
+ head_dim = dim // num_heads
83
+ self.scale = qk_scale or head_dim ** -0.5
84
+
85
+ # define a parameter table of relative position bias
86
+ self.relative_position_bias_table = nn.Parameter(
87
+ torch.zeros((2 * window_size[0] - 1) * (2 * window_size[1] - 1), num_heads)) # 2*Wh-1 * 2*Ww-1, nH
88
+
89
+ # get pair-wise relative position index for each token inside the window
90
+ coords_h = torch.arange(self.window_size[0])
91
+ coords_w = torch.arange(self.window_size[1])
92
+ coords = torch.stack(torch.meshgrid([coords_h, coords_w])) # 2, Wh, Ww
93
+ coords_flatten = torch.flatten(coords, 1) # 2, Wh*Ww
94
+ relative_coords = coords_flatten[:, :, None] - coords_flatten[:, None, :] # 2, Wh*Ww, Wh*Ww
95
+ relative_coords = relative_coords.permute(1, 2, 0).contiguous() # Wh*Ww, Wh*Ww, 2
96
+ relative_coords[:, :, 0] += self.window_size[0] - 1 # shift to start from 0
97
+ relative_coords[:, :, 1] += self.window_size[1] - 1
98
+ relative_coords[:, :, 0] *= 2 * self.window_size[1] - 1
99
+ relative_position_index = relative_coords.sum(-1) # Wh*Ww, Wh*Ww
100
+ self.register_buffer("relative_position_index", relative_position_index)
101
+
102
+ self.qk = nn.Linear(dim, dim * 2, bias=qkv_bias)
103
+ self.attn_drop = nn.Dropout(attn_drop)
104
+ self.proj = nn.Linear(v_dim, v_dim)
105
+ self.proj_drop = nn.Dropout(proj_drop)
106
+
107
+ trunc_normal_(self.relative_position_bias_table, std=.02)
108
+ self.softmax = nn.Softmax(dim=-1)
109
+
110
+ def forward(self, x, v, mask=None):
111
+ """ Forward function.
112
+
113
+ Args:
114
+ x: input features with shape of (num_windows*B, N, C)
115
+ mask: (0/-inf) mask with shape of (num_windows, Wh*Ww, Wh*Ww) or None
116
+ """
117
+ B_, N, C = x.shape
118
+ qk = self.qk(x).reshape(B_, N, 2, self.num_heads, C // self.num_heads).permute(2, 0, 3, 1, 4)
119
+ q, k = qk[0], qk[1] # make torchscript happy (cannot use tensor as tuple)
120
+
121
+ q = q * self.scale
122
+ attn = (q @ k.transpose(-2, -1))
123
+
124
+ relative_position_bias = self.relative_position_bias_table[self.relative_position_index.view(-1)].view(
125
+ self.window_size[0] * self.window_size[1], self.window_size[0] * self.window_size[1], -1) # Wh*Ww,Wh*Ww,nH
126
+ relative_position_bias = relative_position_bias.permute(2, 0, 1).contiguous() # nH, Wh*Ww, Wh*Ww
127
+ attn = attn + relative_position_bias.unsqueeze(0)
128
+
129
+ if mask is not None:
130
+ nW = mask.shape[0]
131
+ attn = attn.view(B_ // nW, nW, self.num_heads, N, N) + mask.unsqueeze(1).unsqueeze(0)
132
+ attn = attn.view(-1, self.num_heads, N, N)
133
+ attn = self.softmax(attn)
134
+ else:
135
+ attn = self.softmax(attn)
136
+
137
+ attn = self.attn_drop(attn)
138
+
139
+ # assert self.dim % v.shape[-1] == 0, "self.dim % v.shape[-1] != 0"
140
+ # repeat_num = self.dim // v.shape[-1]
141
+ # v = v.view(B_, N, self.num_heads // repeat_num, -1).transpose(1, 2).repeat(1, repeat_num, 1, 1)
142
+
143
+ assert self.dim == v.shape[-1], "self.dim != v.shape[-1]"
144
+ v = v.view(B_, N, self.num_heads, -1).transpose(1, 2)
145
+
146
+ x = (attn @ v).transpose(1, 2).reshape(B_, N, C)
147
+ x = self.proj(x)
148
+ x = self.proj_drop(x)
149
+ return x
150
+
151
+
152
+ class CRFBlock(nn.Module):
153
+ """ CRF Block.
154
+
155
+ Args:
156
+ dim (int): Number of input channels.
157
+ num_heads (int): Number of attention heads.
158
+ window_size (int): Window size.
159
+ shift_size (int): Shift size for SW-MSA.
160
+ mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
161
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
162
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
163
+ drop (float, optional): Dropout rate. Default: 0.0
164
+ attn_drop (float, optional): Attention dropout rate. Default: 0.0
165
+ drop_path (float, optional): Stochastic depth rate. Default: 0.0
166
+ act_layer (nn.Module, optional): Activation layer. Default: nn.GELU
167
+ norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
168
+ """
169
+
170
+ def __init__(self, dim, num_heads, v_dim, window_size=7, shift_size=0,
171
+ mlp_ratio=4., qkv_bias=True, qk_scale=None, drop=0., attn_drop=0., drop_path=0.,
172
+ act_layer=nn.GELU, norm_layer=nn.LayerNorm):
173
+ super().__init__()
174
+ self.dim = dim
175
+ self.num_heads = num_heads
176
+ self.v_dim = v_dim
177
+ self.window_size = window_size
178
+ self.shift_size = shift_size
179
+ self.mlp_ratio = mlp_ratio
180
+ assert 0 <= self.shift_size < self.window_size, "shift_size must in 0-window_size"
181
+
182
+ self.norm1 = norm_layer(dim)
183
+ self.attn = WindowAttention(
184
+ dim, window_size=to_2tuple(self.window_size), num_heads=num_heads, v_dim=v_dim,
185
+ qkv_bias=qkv_bias, qk_scale=qk_scale, attn_drop=attn_drop, proj_drop=drop)
186
+
187
+ self.drop_path = DropPath(drop_path) if drop_path > 0. else nn.Identity()
188
+ self.norm2 = norm_layer(v_dim)
189
+ mlp_hidden_dim = int(v_dim * mlp_ratio)
190
+ self.mlp = Mlp(in_features=v_dim, hidden_features=mlp_hidden_dim, act_layer=act_layer, drop=drop)
191
+
192
+ self.H = None
193
+ self.W = None
194
+
195
+ def forward(self, x, v, mask_matrix):
196
+ """ Forward function.
197
+
198
+ Args:
199
+ x: Input feature, tensor size (B, H*W, C).
200
+ H, W: Spatial resolution of the input feature.
201
+ mask_matrix: Attention mask for cyclic shift.
202
+ """
203
+ B, L, C = x.shape
204
+ H, W = self.H, self.W
205
+ assert L == H * W, "input feature has wrong size"
206
+
207
+ shortcut = x
208
+ x = self.norm1(x)
209
+ x = x.view(B, H, W, C)
210
+
211
+ # pad feature maps to multiples of window size
212
+ pad_l = pad_t = 0
213
+ pad_r = (self.window_size - W % self.window_size) % self.window_size
214
+ pad_b = (self.window_size - H % self.window_size) % self.window_size
215
+ x = F.pad(x, (0, 0, pad_l, pad_r, pad_t, pad_b))
216
+ v = F.pad(v, (0, 0, pad_l, pad_r, pad_t, pad_b))
217
+ _, Hp, Wp, _ = x.shape
218
+
219
+ # cyclic shift
220
+ if self.shift_size > 0:
221
+ shifted_x = torch.roll(x, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
222
+ shifted_v = torch.roll(v, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
223
+ attn_mask = mask_matrix
224
+ else:
225
+ shifted_x = x
226
+ shifted_v = v
227
+ attn_mask = None
228
+
229
+ # partition windows
230
+ x_windows = window_partition(shifted_x, self.window_size) # nW*B, window_size, window_size, C
231
+ x_windows = x_windows.view(-1, self.window_size * self.window_size, C) # nW*B, window_size*window_size, C
232
+ v_windows = window_partition(shifted_v, self.window_size) # nW*B, window_size, window_size, C
233
+ v_windows = v_windows.view(-1, self.window_size * self.window_size, v_windows.shape[-1]) # nW*B, window_size*window_size, C
234
+
235
+ # W-MSA/SW-MSA
236
+ attn_windows = self.attn(x_windows, v_windows, mask=attn_mask) # nW*B, window_size*window_size, C
237
+
238
+ # merge windows
239
+ attn_windows = attn_windows.view(-1, self.window_size, self.window_size, self.v_dim)
240
+ shifted_x = window_reverse(attn_windows, self.window_size, Hp, Wp) # B H' W' C
241
+
242
+ # reverse cyclic shift
243
+ if self.shift_size > 0:
244
+ x = torch.roll(shifted_x, shifts=(self.shift_size, self.shift_size), dims=(1, 2))
245
+ else:
246
+ x = shifted_x
247
+
248
+ if pad_r > 0 or pad_b > 0:
249
+ x = x[:, :H, :W, :].contiguous()
250
+
251
+ x = x.view(B, H * W, self.v_dim)
252
+
253
+ # FFN
254
+ x = shortcut + self.drop_path(x)
255
+ x = x + self.drop_path(self.mlp(self.norm2(x)))
256
+
257
+ return x
258
+
259
+
260
+ class BasicCRFLayer(nn.Module):
261
+ """ A basic NeWCRFs layer for one stage.
262
+
263
+ Args:
264
+ dim (int): Number of feature channels
265
+ depth (int): Depths of this stage.
266
+ num_heads (int): Number of attention head.
267
+ window_size (int): Local window size. Default: 7.
268
+ mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
269
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
270
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
271
+ drop (float, optional): Dropout rate. Default: 0.0
272
+ attn_drop (float, optional): Attention dropout rate. Default: 0.0
273
+ drop_path (float | tuple[float], optional): Stochastic depth rate. Default: 0.0
274
+ norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
275
+ downsample (nn.Module | None, optional): Downsample layer at the end of the layer. Default: None
276
+ use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
277
+ """
278
+
279
+ def __init__(self,
280
+ dim,
281
+ depth,
282
+ num_heads,
283
+ v_dim,
284
+ window_size=7,
285
+ mlp_ratio=4.,
286
+ qkv_bias=True,
287
+ qk_scale=None,
288
+ drop=0.,
289
+ attn_drop=0.,
290
+ drop_path=0.,
291
+ norm_layer=nn.LayerNorm,
292
+ downsample=None,
293
+ use_checkpoint=False):
294
+ super().__init__()
295
+ self.window_size = window_size
296
+ self.shift_size = window_size // 2
297
+ self.depth = depth
298
+ self.use_checkpoint = use_checkpoint
299
+
300
+ # build blocks
301
+ self.blocks = nn.ModuleList([
302
+ CRFBlock(
303
+ dim=dim,
304
+ num_heads=num_heads,
305
+ v_dim=v_dim,
306
+ window_size=window_size,
307
+ shift_size=0 if (i % 2 == 0) else window_size // 2,
308
+ mlp_ratio=mlp_ratio,
309
+ qkv_bias=qkv_bias,
310
+ qk_scale=qk_scale,
311
+ drop=drop,
312
+ attn_drop=attn_drop,
313
+ drop_path=drop_path[i] if isinstance(drop_path, list) else drop_path,
314
+ norm_layer=norm_layer)
315
+ for i in range(depth)])
316
+
317
+ # patch merging layer
318
+ if downsample is not None:
319
+ self.downsample = downsample(dim=dim, norm_layer=norm_layer)
320
+ else:
321
+ self.downsample = None
322
+
323
+ def forward(self, x, v, H, W):
324
+ """ Forward function.
325
+
326
+ Args:
327
+ x: Input feature, tensor size (B, H*W, C).
328
+ H, W: Spatial resolution of the input feature.
329
+ """
330
+
331
+ # calculate attention mask for SW-MSA
332
+ Hp = int(np.ceil(H / self.window_size)) * self.window_size
333
+ Wp = int(np.ceil(W / self.window_size)) * self.window_size
334
+ img_mask = torch.zeros((1, Hp, Wp, 1), device=x.device) # 1 Hp Wp 1
335
+ h_slices = (slice(0, -self.window_size),
336
+ slice(-self.window_size, -self.shift_size),
337
+ slice(-self.shift_size, None))
338
+ w_slices = (slice(0, -self.window_size),
339
+ slice(-self.window_size, -self.shift_size),
340
+ slice(-self.shift_size, None))
341
+ cnt = 0
342
+ for h in h_slices:
343
+ for w in w_slices:
344
+ img_mask[:, h, w, :] = cnt
345
+ cnt += 1
346
+
347
+ mask_windows = window_partition(img_mask, self.window_size) # nW, window_size, window_size, 1
348
+ mask_windows = mask_windows.view(-1, self.window_size * self.window_size)
349
+ attn_mask = mask_windows.unsqueeze(1) - mask_windows.unsqueeze(2)
350
+ attn_mask = attn_mask.masked_fill(attn_mask != 0, float(-100.0)).masked_fill(attn_mask == 0, float(0.0))
351
+
352
+ for blk in self.blocks:
353
+ blk.H, blk.W = H, W
354
+ if self.use_checkpoint:
355
+ x = checkpoint.checkpoint(blk, x, attn_mask)
356
+ else:
357
+ x = blk(x, v, attn_mask)
358
+ if self.downsample is not None:
359
+ x_down = self.downsample(x, H, W)
360
+ Wh, Ww = (H + 1) // 2, (W + 1) // 2
361
+ return x, H, W, x_down, Wh, Ww
362
+ else:
363
+ return x, H, W, x, H, W
364
+
365
+
366
+ class NewCRF(nn.Module):
367
+ def __init__(self,
368
+ input_dim=96,
369
+ embed_dim=96,
370
+ v_dim=64,
371
+ window_size=7,
372
+ num_heads=4,
373
+ depth=2,
374
+ patch_size=4,
375
+ in_chans=3,
376
+ norm_layer=nn.LayerNorm,
377
+ patch_norm=True):
378
+ super().__init__()
379
+
380
+ self.embed_dim = embed_dim
381
+ self.patch_norm = patch_norm
382
+
383
+ if input_dim != embed_dim:
384
+ self.proj_x = nn.Conv2d(input_dim, embed_dim, 3, padding=1)
385
+ else:
386
+ self.proj_x = None
387
+
388
+ if v_dim != embed_dim:
389
+ self.proj_v = nn.Conv2d(v_dim, embed_dim, 3, padding=1)
390
+ elif embed_dim % v_dim == 0:
391
+ self.proj_v = None
392
+
393
+ # For now, v_dim need to be equal to embed_dim, because the output of window-attn is the input of shift-window-attn
394
+ v_dim = embed_dim
395
+ assert v_dim == embed_dim
396
+
397
+ self.crf_layer = BasicCRFLayer(
398
+ dim=embed_dim,
399
+ depth=depth,
400
+ num_heads=num_heads,
401
+ v_dim=v_dim,
402
+ window_size=window_size,
403
+ mlp_ratio=4.,
404
+ qkv_bias=True,
405
+ qk_scale=None,
406
+ drop=0.,
407
+ attn_drop=0.,
408
+ drop_path=0.,
409
+ norm_layer=norm_layer,
410
+ downsample=None,
411
+ use_checkpoint=False)
412
+
413
+ layer = norm_layer(embed_dim)
414
+ layer_name = 'norm_crf'
415
+ self.add_module(layer_name, layer)
416
+
417
+
418
+ def forward(self, x, v):
419
+ if self.proj_x is not None:
420
+ x = self.proj_x(x)
421
+ if self.proj_v is not None:
422
+ v = self.proj_v(v)
423
+
424
+ Wh, Ww = x.size(2), x.size(3)
425
+ x = x.flatten(2).transpose(1, 2)
426
+ v = v.transpose(1, 2).transpose(2, 3)
427
+
428
+ x_out, H, W, x, Wh, Ww = self.crf_layer(x, v, Wh, Ww)
429
+ norm_layer = getattr(self, f'norm_crf')
430
+ x_out = norm_layer(x_out)
431
+ out = x_out.view(-1, H, W, self.embed_dim).permute(0, 3, 1, 2).contiguous()
432
+
433
+ return out
external/WildCamera/WildCamera/newcrfs/newcrf_utils.py ADDED
@@ -0,0 +1,264 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import warnings
2
+ import os
3
+ import os.path as osp
4
+ import pkgutil
5
+ import warnings
6
+ from collections import OrderedDict
7
+ from importlib import import_module
8
+
9
+ import torch
10
+ import torchvision
11
+ import torch.nn as nn
12
+ from torch.utils import model_zoo
13
+ from torch.nn import functional as F
14
+ from torch.nn.parallel import DataParallel, DistributedDataParallel
15
+ from torch import distributed as dist
16
+
17
+ TORCH_VERSION = torch.__version__
18
+
19
+
20
+ def resize(input,
21
+ size=None,
22
+ scale_factor=None,
23
+ mode='nearest',
24
+ align_corners=None,
25
+ warning=True):
26
+ if warning:
27
+ if size is not None and align_corners:
28
+ input_h, input_w = tuple(int(x) for x in input.shape[2:])
29
+ output_h, output_w = tuple(int(x) for x in size)
30
+ if output_h > input_h or output_w > output_h:
31
+ if ((output_h > 1 and output_w > 1 and input_h > 1
32
+ and input_w > 1) and (output_h - 1) % (input_h - 1)
33
+ and (output_w - 1) % (input_w - 1)):
34
+ warnings.warn(
35
+ f'When align_corners={align_corners}, '
36
+ 'the output would more aligned if '
37
+ f'input size {(input_h, input_w)} is `x+1` and '
38
+ f'out size {(output_h, output_w)} is `nx+1`')
39
+ if isinstance(size, torch.Size):
40
+ size = tuple(int(x) for x in size)
41
+ return F.interpolate(input, size, scale_factor, mode, align_corners)
42
+
43
+
44
+ def normal_init(module, mean=0, std=1, bias=0):
45
+ if hasattr(module, 'weight') and module.weight is not None:
46
+ nn.init.normal_(module.weight, mean, std)
47
+ if hasattr(module, 'bias') and module.bias is not None:
48
+ nn.init.constant_(module.bias, bias)
49
+
50
+
51
+ def is_module_wrapper(module):
52
+ module_wrappers = (DataParallel, DistributedDataParallel)
53
+ return isinstance(module, module_wrappers)
54
+
55
+
56
+ def get_dist_info():
57
+ if TORCH_VERSION < '1.0':
58
+ initialized = dist._initialized
59
+ else:
60
+ if dist.is_available():
61
+ initialized = dist.is_initialized()
62
+ else:
63
+ initialized = False
64
+ if initialized:
65
+ rank = dist.get_rank()
66
+ world_size = dist.get_world_size()
67
+ else:
68
+ rank = 0
69
+ world_size = 1
70
+ return rank, world_size
71
+
72
+
73
+ def load_state_dict(module, state_dict, strict=False, logger=None):
74
+ """Load state_dict to a module.
75
+
76
+ This method is modified from :meth:`torch.nn.Module.load_state_dict`.
77
+ Default value for ``strict`` is set to ``False`` and the message for
78
+ param mismatch will be shown even if strict is False.
79
+
80
+ Args:
81
+ module (Module): Module that receives the state_dict.
82
+ state_dict (OrderedDict): Weights.
83
+ strict (bool): whether to strictly enforce that the keys
84
+ in :attr:`state_dict` match the keys returned by this module's
85
+ :meth:`~torch.nn.Module.state_dict` function. Default: ``False``.
86
+ logger (:obj:`logging.Logger`, optional): Logger to log the error
87
+ message. If not specified, print function will be used.
88
+ """
89
+ unexpected_keys = []
90
+ all_missing_keys = []
91
+ err_msg = []
92
+
93
+ metadata = getattr(state_dict, '_metadata', None)
94
+ state_dict = state_dict.copy()
95
+ if metadata is not None:
96
+ state_dict._metadata = metadata
97
+
98
+ # use _load_from_state_dict to enable checkpoint version control
99
+ def load(module, prefix=''):
100
+ # recursively check parallel module in case that the model has a
101
+ # complicated structure, e.g., nn.Module(nn.Module(DDP))
102
+ if is_module_wrapper(module):
103
+ module = module.module
104
+ local_metadata = {} if metadata is None else metadata.get(
105
+ prefix[:-1], {})
106
+ module._load_from_state_dict(state_dict, prefix, local_metadata, True,
107
+ all_missing_keys, unexpected_keys,
108
+ err_msg)
109
+ for name, child in module._modules.items():
110
+ if child is not None:
111
+ load(child, prefix + name + '.')
112
+
113
+ load(module)
114
+ load = None # break load->load reference cycle
115
+
116
+ # ignore "num_batches_tracked" of BN layers
117
+ missing_keys = [
118
+ key for key in all_missing_keys if 'num_batches_tracked' not in key
119
+ ]
120
+
121
+ if unexpected_keys:
122
+ err_msg.append('unexpected key in source '
123
+ f'state_dict: {", ".join(unexpected_keys)}\n')
124
+ if missing_keys:
125
+ err_msg.append(
126
+ f'missing keys in source state_dict: {", ".join(missing_keys)}\n')
127
+
128
+ rank, _ = get_dist_info()
129
+ if len(err_msg) > 0 and rank == 0:
130
+ err_msg.insert(
131
+ 0, 'The model and loaded state dict do not match exactly\n')
132
+ err_msg = '\n'.join(err_msg)
133
+ if strict:
134
+ raise RuntimeError(err_msg)
135
+ elif logger is not None:
136
+ logger.warning(err_msg)
137
+ else:
138
+ print(err_msg)
139
+
140
+
141
+ def load_url_dist(url, model_dir=None):
142
+ """In distributed setting, this function only download checkpoint at local
143
+ rank 0."""
144
+ rank, world_size = get_dist_info()
145
+ rank = int(os.environ.get('LOCAL_RANK', rank))
146
+ if rank == 0:
147
+ checkpoint = model_zoo.load_url(url, model_dir=model_dir)
148
+ if world_size > 1:
149
+ torch.distributed.barrier()
150
+ if rank > 0:
151
+ checkpoint = model_zoo.load_url(url, model_dir=model_dir)
152
+ return checkpoint
153
+
154
+
155
+ def get_torchvision_models():
156
+ model_urls = dict()
157
+ for _, name, ispkg in pkgutil.walk_packages(torchvision.models.__path__):
158
+ if ispkg:
159
+ continue
160
+ _zoo = import_module(f'torchvision.models.{name}')
161
+ if hasattr(_zoo, 'model_urls'):
162
+ _urls = getattr(_zoo, 'model_urls')
163
+ model_urls.update(_urls)
164
+ return model_urls
165
+
166
+
167
+ def _load_checkpoint(filename, map_location=None):
168
+ """Load checkpoint from somewhere (modelzoo, file, url).
169
+
170
+ Args:
171
+ filename (str): Accept local filepath, URL, ``torchvision://xxx``,
172
+ ``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for
173
+ details.
174
+ map_location (str | None): Same as :func:`torch.load`. Default: None.
175
+
176
+ Returns:
177
+ dict | OrderedDict: The loaded checkpoint. It can be either an
178
+ OrderedDict storing model weights or a dict containing other
179
+ information, which depends on the checkpoint.
180
+ """
181
+ if filename.startswith('modelzoo://'):
182
+ warnings.warn('The URL scheme of "modelzoo://" is deprecated, please '
183
+ 'use "torchvision://" instead')
184
+ model_urls = get_torchvision_models()
185
+ model_name = filename[11:]
186
+ checkpoint = load_url_dist(model_urls[model_name])
187
+ else:
188
+ if not osp.isfile(filename):
189
+ raise IOError(f'{filename} is not a checkpoint file')
190
+ checkpoint = torch.load(filename, map_location=map_location)
191
+ return checkpoint
192
+
193
+
194
+ def load_checkpoint(model,
195
+ filename,
196
+ map_location='cpu',
197
+ strict=False,
198
+ logger=None):
199
+ """Load checkpoint from a file or URI.
200
+
201
+ Args:
202
+ model (Module): Module to load checkpoint.
203
+ filename (str): Accept local filepath, URL, ``torchvision://xxx``,
204
+ ``open-mmlab://xxx``. Please refer to ``docs/model_zoo.md`` for
205
+ details.
206
+ map_location (str): Same as :func:`torch.load`.
207
+ strict (bool): Whether to allow different params for the model and
208
+ checkpoint.
209
+ logger (:mod:`logging.Logger` or None): The logger for error message.
210
+
211
+ Returns:
212
+ dict or OrderedDict: The loaded checkpoint.
213
+ """
214
+ checkpoint = _load_checkpoint(filename, map_location)
215
+ # OrderedDict is a subclass of dict
216
+ if not isinstance(checkpoint, dict):
217
+ raise RuntimeError(
218
+ f'No state_dict found in checkpoint file {filename}')
219
+ # get state_dict from checkpoint
220
+ if 'state_dict' in checkpoint:
221
+ state_dict = checkpoint['state_dict']
222
+ elif 'model' in checkpoint:
223
+ state_dict = checkpoint['model']
224
+ else:
225
+ state_dict = checkpoint
226
+ # strip prefix of state_dict
227
+ if list(state_dict.keys())[0].startswith('module.'):
228
+ state_dict = {k[7:]: v for k, v in state_dict.items()}
229
+
230
+ # for MoBY, load model of online branch
231
+ if sorted(list(state_dict.keys()))[0].startswith('encoder'):
232
+ state_dict = {k.replace('encoder.', ''): v for k, v in state_dict.items() if k.startswith('encoder.')}
233
+
234
+ # reshape absolute position embedding
235
+ if state_dict.get('absolute_pos_embed') is not None:
236
+ absolute_pos_embed = state_dict['absolute_pos_embed']
237
+ N1, L, C1 = absolute_pos_embed.size()
238
+ N2, C2, H, W = model.absolute_pos_embed.size()
239
+ if N1 != N2 or C1 != C2 or L != H*W:
240
+ logger.warning("Error in loading absolute_pos_embed, pass")
241
+ else:
242
+ state_dict['absolute_pos_embed'] = absolute_pos_embed.view(N2, H, W, C2).permute(0, 3, 1, 2)
243
+
244
+ # interpolate position bias table if needed
245
+ relative_position_bias_table_keys = [k for k in state_dict.keys() if "relative_position_bias_table" in k]
246
+ for table_key in relative_position_bias_table_keys:
247
+ table_pretrained = state_dict[table_key]
248
+ table_current = model.state_dict()[table_key]
249
+ L1, nH1 = table_pretrained.size()
250
+ L2, nH2 = table_current.size()
251
+ if nH1 != nH2:
252
+ logger.warning(f"Error in loading {table_key}, pass")
253
+ else:
254
+ if L1 != L2:
255
+ S1 = int(L1 ** 0.5)
256
+ S2 = int(L2 ** 0.5)
257
+ table_pretrained_resized = F.interpolate(
258
+ table_pretrained.permute(1, 0).view(1, nH1, S1, S1),
259
+ size=(S2, S2), mode='bicubic')
260
+ state_dict[table_key] = table_pretrained_resized.view(nH2, L2).permute(1, 0)
261
+
262
+ # load state_dict
263
+ load_state_dict(model, state_dict, strict, logger)
264
+ return checkpoint
external/WildCamera/WildCamera/newcrfs/swin_transformer.py ADDED
@@ -0,0 +1,619 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+ import torch.nn.functional as F
4
+ import torch.utils.checkpoint as checkpoint
5
+ import numpy as np
6
+ from timm.models.layers import DropPath, to_2tuple, trunc_normal_
7
+
8
+ from .newcrf_utils import load_checkpoint
9
+
10
+ class Mlp(nn.Module):
11
+ """ Multilayer perceptron."""
12
+
13
+ def __init__(self, in_features, hidden_features=None, out_features=None, act_layer=nn.GELU, drop=0.):
14
+ super().__init__()
15
+ out_features = out_features or in_features
16
+ hidden_features = hidden_features or in_features
17
+ self.fc1 = nn.Linear(in_features, hidden_features)
18
+ self.act = act_layer()
19
+ self.fc2 = nn.Linear(hidden_features, out_features)
20
+ self.drop = nn.Dropout(drop)
21
+
22
+ def forward(self, x):
23
+ x = self.fc1(x)
24
+ x = self.act(x)
25
+ x = self.drop(x)
26
+ x = self.fc2(x)
27
+ x = self.drop(x)
28
+ return x
29
+
30
+
31
+ def window_partition(x, window_size):
32
+ """
33
+ Args:
34
+ x: (B, H, W, C)
35
+ window_size (int): window size
36
+
37
+ Returns:
38
+ windows: (num_windows*B, window_size, window_size, C)
39
+ """
40
+ B, H, W, C = x.shape
41
+ x = x.view(B, H // window_size, window_size, W // window_size, window_size, C)
42
+ windows = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(-1, window_size, window_size, C)
43
+ return windows
44
+
45
+
46
+ def window_reverse(windows, window_size, H, W):
47
+ """
48
+ Args:
49
+ windows: (num_windows*B, window_size, window_size, C)
50
+ window_size (int): Window size
51
+ H (int): Height of image
52
+ W (int): Width of image
53
+
54
+ Returns:
55
+ x: (B, H, W, C)
56
+ """
57
+ B = int(windows.shape[0] / (H * W / window_size / window_size))
58
+ x = windows.view(B, H // window_size, W // window_size, window_size, window_size, -1)
59
+ x = x.permute(0, 1, 3, 2, 4, 5).contiguous().view(B, H, W, -1)
60
+ return x
61
+
62
+
63
+ class WindowAttention(nn.Module):
64
+ """ Window based multi-head self attention (W-MSA) module with relative position bias.
65
+ It supports both of shifted and non-shifted window.
66
+
67
+ Args:
68
+ dim (int): Number of input channels.
69
+ window_size (tuple[int]): The height and width of the window.
70
+ num_heads (int): Number of attention heads.
71
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
72
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set
73
+ attn_drop (float, optional): Dropout ratio of attention weight. Default: 0.0
74
+ proj_drop (float, optional): Dropout ratio of output. Default: 0.0
75
+ """
76
+
77
+ def __init__(self, dim, window_size, num_heads, qkv_bias=True, qk_scale=None, attn_drop=0., proj_drop=0.):
78
+
79
+ super().__init__()
80
+ self.dim = dim
81
+ self.window_size = window_size # Wh, Ww
82
+ self.num_heads = num_heads
83
+ head_dim = dim // num_heads
84
+ self.scale = qk_scale or head_dim ** -0.5
85
+
86
+ # define a parameter table of relative position bias
87
+ self.relative_position_bias_table = nn.Parameter(
88
+ torch.zeros((2 * window_size[0] - 1) * (2 * window_size[1] - 1), num_heads)) # 2*Wh-1 * 2*Ww-1, nH
89
+
90
+ # get pair-wise relative position index for each token inside the window
91
+ coords_h = torch.arange(self.window_size[0])
92
+ coords_w = torch.arange(self.window_size[1])
93
+ coords = torch.stack(torch.meshgrid([coords_h, coords_w])) # 2, Wh, Ww
94
+ coords_flatten = torch.flatten(coords, 1) # 2, Wh*Ww
95
+ relative_coords = coords_flatten[:, :, None] - coords_flatten[:, None, :] # 2, Wh*Ww, Wh*Ww
96
+ relative_coords = relative_coords.permute(1, 2, 0).contiguous() # Wh*Ww, Wh*Ww, 2
97
+ relative_coords[:, :, 0] += self.window_size[0] - 1 # shift to start from 0
98
+ relative_coords[:, :, 1] += self.window_size[1] - 1
99
+ relative_coords[:, :, 0] *= 2 * self.window_size[1] - 1
100
+ relative_position_index = relative_coords.sum(-1) # Wh*Ww, Wh*Ww
101
+ self.register_buffer("relative_position_index", relative_position_index)
102
+
103
+ self.qkv = nn.Linear(dim, dim * 3, bias=qkv_bias)
104
+ self.attn_drop = nn.Dropout(attn_drop)
105
+ self.proj = nn.Linear(dim, dim)
106
+ self.proj_drop = nn.Dropout(proj_drop)
107
+
108
+ trunc_normal_(self.relative_position_bias_table, std=.02)
109
+ self.softmax = nn.Softmax(dim=-1)
110
+
111
+ def forward(self, x, mask=None):
112
+ """ Forward function.
113
+
114
+ Args:
115
+ x: input features with shape of (num_windows*B, N, C)
116
+ mask: (0/-inf) mask with shape of (num_windows, Wh*Ww, Wh*Ww) or None
117
+ """
118
+ B_, N, C = x.shape
119
+ qkv = self.qkv(x).reshape(B_, N, 3, self.num_heads, C // self.num_heads).permute(2, 0, 3, 1, 4)
120
+ q, k, v = qkv[0], qkv[1], qkv[2] # make torchscript happy (cannot use tensor as tuple)
121
+
122
+ q = q * self.scale
123
+ attn = (q @ k.transpose(-2, -1))
124
+
125
+ relative_position_bias = self.relative_position_bias_table[self.relative_position_index.view(-1)].view(
126
+ self.window_size[0] * self.window_size[1], self.window_size[0] * self.window_size[1], -1) # Wh*Ww,Wh*Ww,nH
127
+ relative_position_bias = relative_position_bias.permute(2, 0, 1).contiguous() # nH, Wh*Ww, Wh*Ww
128
+ attn = attn + relative_position_bias.unsqueeze(0)
129
+
130
+ if mask is not None:
131
+ nW = mask.shape[0]
132
+ attn = attn.view(B_ // nW, nW, self.num_heads, N, N) + mask.unsqueeze(1).unsqueeze(0)
133
+ attn = attn.view(-1, self.num_heads, N, N)
134
+ attn = self.softmax(attn)
135
+ else:
136
+ attn = self.softmax(attn)
137
+
138
+ attn = self.attn_drop(attn)
139
+
140
+ x = (attn @ v).transpose(1, 2).reshape(B_, N, C)
141
+ x = self.proj(x)
142
+ x = self.proj_drop(x)
143
+ return x
144
+
145
+
146
+ class SwinTransformerBlock(nn.Module):
147
+ """ Swin Transformer Block.
148
+
149
+ Args:
150
+ dim (int): Number of input channels.
151
+ num_heads (int): Number of attention heads.
152
+ window_size (int): Window size.
153
+ shift_size (int): Shift size for SW-MSA.
154
+ mlp_ratio (float): Ratio of mlp hidden dim to embedding dim.
155
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
156
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
157
+ drop (float, optional): Dropout rate. Default: 0.0
158
+ attn_drop (float, optional): Attention dropout rate. Default: 0.0
159
+ drop_path (float, optional): Stochastic depth rate. Default: 0.0
160
+ act_layer (nn.Module, optional): Activation layer. Default: nn.GELU
161
+ norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
162
+ """
163
+
164
+ def __init__(self, dim, num_heads, window_size=7, shift_size=0,
165
+ mlp_ratio=4., qkv_bias=True, qk_scale=None, drop=0., attn_drop=0., drop_path=0.,
166
+ act_layer=nn.GELU, norm_layer=nn.LayerNorm):
167
+ super().__init__()
168
+ self.dim = dim
169
+ self.num_heads = num_heads
170
+ self.window_size = window_size
171
+ self.shift_size = shift_size
172
+ self.mlp_ratio = mlp_ratio
173
+ assert 0 <= self.shift_size < self.window_size, "shift_size must in 0-window_size"
174
+
175
+ self.norm1 = norm_layer(dim)
176
+ self.attn = WindowAttention(
177
+ dim, window_size=to_2tuple(self.window_size), num_heads=num_heads,
178
+ qkv_bias=qkv_bias, qk_scale=qk_scale, attn_drop=attn_drop, proj_drop=drop)
179
+
180
+ self.drop_path = DropPath(drop_path) if drop_path > 0. else nn.Identity()
181
+ self.norm2 = norm_layer(dim)
182
+ mlp_hidden_dim = int(dim * mlp_ratio)
183
+ self.mlp = Mlp(in_features=dim, hidden_features=mlp_hidden_dim, act_layer=act_layer, drop=drop)
184
+
185
+ self.H = None
186
+ self.W = None
187
+
188
+ def forward(self, x, mask_matrix):
189
+ """ Forward function.
190
+
191
+ Args:
192
+ x: Input feature, tensor size (B, H*W, C).
193
+ H, W: Spatial resolution of the input feature.
194
+ mask_matrix: Attention mask for cyclic shift.
195
+ """
196
+ B, L, C = x.shape
197
+ H, W = self.H, self.W
198
+ assert L == H * W, "input feature has wrong size"
199
+
200
+ shortcut = x
201
+ x = self.norm1(x)
202
+ x = x.view(B, H, W, C)
203
+
204
+ # pad feature maps to multiples of window size
205
+ pad_l = pad_t = 0
206
+ pad_r = (self.window_size - W % self.window_size) % self.window_size
207
+ pad_b = (self.window_size - H % self.window_size) % self.window_size
208
+ x = F.pad(x, (0, 0, pad_l, pad_r, pad_t, pad_b))
209
+ _, Hp, Wp, _ = x.shape
210
+
211
+ # cyclic shift
212
+ if self.shift_size > 0:
213
+ shifted_x = torch.roll(x, shifts=(-self.shift_size, -self.shift_size), dims=(1, 2))
214
+ attn_mask = mask_matrix
215
+ else:
216
+ shifted_x = x
217
+ attn_mask = None
218
+
219
+ # partition windows
220
+ x_windows = window_partition(shifted_x, self.window_size) # nW*B, window_size, window_size, C
221
+ x_windows = x_windows.view(-1, self.window_size * self.window_size, C) # nW*B, window_size*window_size, C
222
+
223
+ # W-MSA/SW-MSA
224
+ attn_windows = self.attn(x_windows, mask=attn_mask) # nW*B, window_size*window_size, C
225
+
226
+ # merge windows
227
+ attn_windows = attn_windows.view(-1, self.window_size, self.window_size, C)
228
+ shifted_x = window_reverse(attn_windows, self.window_size, Hp, Wp) # B H' W' C
229
+
230
+ # reverse cyclic shift
231
+ if self.shift_size > 0:
232
+ x = torch.roll(shifted_x, shifts=(self.shift_size, self.shift_size), dims=(1, 2))
233
+ else:
234
+ x = shifted_x
235
+
236
+ if pad_r > 0 or pad_b > 0:
237
+ x = x[:, :H, :W, :].contiguous()
238
+
239
+ x = x.view(B, H * W, C)
240
+
241
+ # FFN
242
+ x = shortcut + self.drop_path(x)
243
+ x = x + self.drop_path(self.mlp(self.norm2(x)))
244
+
245
+ return x
246
+
247
+
248
+ class PatchMerging(nn.Module):
249
+ """ Patch Merging Layer
250
+
251
+ Args:
252
+ dim (int): Number of input channels.
253
+ norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
254
+ """
255
+ def __init__(self, dim, norm_layer=nn.LayerNorm):
256
+ super().__init__()
257
+ self.dim = dim
258
+ self.reduction = nn.Linear(4 * dim, 2 * dim, bias=False)
259
+ self.norm = norm_layer(4 * dim)
260
+
261
+ def forward(self, x, H, W):
262
+ """ Forward function.
263
+
264
+ Args:
265
+ x: Input feature, tensor size (B, H*W, C).
266
+ H, W: Spatial resolution of the input feature.
267
+ """
268
+ B, L, C = x.shape
269
+ assert L == H * W, "input feature has wrong size"
270
+
271
+ x = x.view(B, H, W, C)
272
+
273
+ # padding
274
+ pad_input = (H % 2 == 1) or (W % 2 == 1)
275
+ if pad_input:
276
+ x = F.pad(x, (0, 0, 0, W % 2, 0, H % 2))
277
+
278
+ x0 = x[:, 0::2, 0::2, :] # B H/2 W/2 C
279
+ x1 = x[:, 1::2, 0::2, :] # B H/2 W/2 C
280
+ x2 = x[:, 0::2, 1::2, :] # B H/2 W/2 C
281
+ x3 = x[:, 1::2, 1::2, :] # B H/2 W/2 C
282
+ x = torch.cat([x0, x1, x2, x3], -1) # B H/2 W/2 4*C
283
+ x = x.view(B, -1, 4 * C) # B H/2*W/2 4*C
284
+
285
+ x = self.norm(x)
286
+ x = self.reduction(x)
287
+
288
+ return x
289
+
290
+
291
+ class BasicLayer(nn.Module):
292
+ """ A basic Swin Transformer layer for one stage.
293
+
294
+ Args:
295
+ dim (int): Number of feature channels
296
+ depth (int): Depths of this stage.
297
+ num_heads (int): Number of attention head.
298
+ window_size (int): Local window size. Default: 7.
299
+ mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
300
+ qkv_bias (bool, optional): If True, add a learnable bias to query, key, value. Default: True
301
+ qk_scale (float | None, optional): Override default qk scale of head_dim ** -0.5 if set.
302
+ drop (float, optional): Dropout rate. Default: 0.0
303
+ attn_drop (float, optional): Attention dropout rate. Default: 0.0
304
+ drop_path (float | tuple[float], optional): Stochastic depth rate. Default: 0.0
305
+ norm_layer (nn.Module, optional): Normalization layer. Default: nn.LayerNorm
306
+ downsample (nn.Module | None, optional): Downsample layer at the end of the layer. Default: None
307
+ use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
308
+ """
309
+
310
+ def __init__(self,
311
+ dim,
312
+ depth,
313
+ num_heads,
314
+ window_size=7,
315
+ mlp_ratio=4.,
316
+ qkv_bias=True,
317
+ qk_scale=None,
318
+ drop=0.,
319
+ attn_drop=0.,
320
+ drop_path=0.,
321
+ norm_layer=nn.LayerNorm,
322
+ downsample=None,
323
+ use_checkpoint=False):
324
+ super().__init__()
325
+ self.window_size = window_size
326
+ self.shift_size = window_size // 2
327
+ self.depth = depth
328
+ self.use_checkpoint = use_checkpoint
329
+
330
+ # build blocks
331
+ self.blocks = nn.ModuleList([
332
+ SwinTransformerBlock(
333
+ dim=dim,
334
+ num_heads=num_heads,
335
+ window_size=window_size,
336
+ shift_size=0 if (i % 2 == 0) else window_size // 2,
337
+ mlp_ratio=mlp_ratio,
338
+ qkv_bias=qkv_bias,
339
+ qk_scale=qk_scale,
340
+ drop=drop,
341
+ attn_drop=attn_drop,
342
+ drop_path=drop_path[i] if isinstance(drop_path, list) else drop_path,
343
+ norm_layer=norm_layer)
344
+ for i in range(depth)])
345
+
346
+ # patch merging layer
347
+ if downsample is not None:
348
+ self.downsample = downsample(dim=dim, norm_layer=norm_layer)
349
+ else:
350
+ self.downsample = None
351
+
352
+ def forward(self, x, H, W):
353
+ """ Forward function.
354
+
355
+ Args:
356
+ x: Input feature, tensor size (B, H*W, C).
357
+ H, W: Spatial resolution of the input feature.
358
+ """
359
+
360
+ # calculate attention mask for SW-MSA
361
+ Hp = int(np.ceil(H / self.window_size)) * self.window_size
362
+ Wp = int(np.ceil(W / self.window_size)) * self.window_size
363
+ img_mask = torch.zeros((1, Hp, Wp, 1), device=x.device) # 1 Hp Wp 1
364
+ h_slices = (slice(0, -self.window_size),
365
+ slice(-self.window_size, -self.shift_size),
366
+ slice(-self.shift_size, None))
367
+ w_slices = (slice(0, -self.window_size),
368
+ slice(-self.window_size, -self.shift_size),
369
+ slice(-self.shift_size, None))
370
+ cnt = 0
371
+ for h in h_slices:
372
+ for w in w_slices:
373
+ img_mask[:, h, w, :] = cnt
374
+ cnt += 1
375
+
376
+ mask_windows = window_partition(img_mask, self.window_size) # nW, window_size, window_size, 1
377
+ mask_windows = mask_windows.view(-1, self.window_size * self.window_size)
378
+ attn_mask = mask_windows.unsqueeze(1) - mask_windows.unsqueeze(2)
379
+ attn_mask = attn_mask.masked_fill(attn_mask != 0, float(-100.0)).masked_fill(attn_mask == 0, float(0.0))
380
+
381
+ for blk in self.blocks:
382
+ blk.H, blk.W = H, W
383
+ if self.use_checkpoint:
384
+ x = checkpoint.checkpoint(blk, x, attn_mask)
385
+ else:
386
+ x = blk(x, attn_mask)
387
+ if self.downsample is not None:
388
+ x_down = self.downsample(x, H, W)
389
+ Wh, Ww = (H + 1) // 2, (W + 1) // 2
390
+ return x, H, W, x_down, Wh, Ww
391
+ else:
392
+ return x, H, W, x, H, W
393
+
394
+
395
+ class PatchEmbed(nn.Module):
396
+ """ Image to Patch Embedding
397
+
398
+ Args:
399
+ patch_size (int): Patch token size. Default: 4.
400
+ in_chans (int): Number of input image channels. Default: 3.
401
+ embed_dim (int): Number of linear projection output channels. Default: 96.
402
+ norm_layer (nn.Module, optional): Normalization layer. Default: None
403
+ """
404
+
405
+ def __init__(self, patch_size=4, in_chans=3, embed_dim=96, norm_layer=None):
406
+ super().__init__()
407
+ patch_size = to_2tuple(patch_size)
408
+ self.patch_size = patch_size
409
+
410
+ self.in_chans = in_chans
411
+ self.embed_dim = embed_dim
412
+
413
+ self.proj = nn.Conv2d(in_chans, embed_dim, kernel_size=patch_size, stride=patch_size)
414
+ if norm_layer is not None:
415
+ self.norm = norm_layer(embed_dim)
416
+ else:
417
+ self.norm = None
418
+
419
+ def forward(self, x):
420
+ """Forward function."""
421
+ # padding
422
+ _, _, H, W = x.size()
423
+ if W % self.patch_size[1] != 0:
424
+ x = F.pad(x, (0, self.patch_size[1] - W % self.patch_size[1]))
425
+ if H % self.patch_size[0] != 0:
426
+ x = F.pad(x, (0, 0, 0, self.patch_size[0] - H % self.patch_size[0]))
427
+
428
+ x = self.proj(x) # B C Wh Ww
429
+ if self.norm is not None:
430
+ Wh, Ww = x.size(2), x.size(3)
431
+ x = x.flatten(2).transpose(1, 2)
432
+ x = self.norm(x)
433
+ x = x.transpose(1, 2).view(-1, self.embed_dim, Wh, Ww)
434
+
435
+ return x
436
+
437
+
438
+ class SwinTransformer(nn.Module):
439
+ """ Swin Transformer backbone.
440
+ A PyTorch impl of : `Swin Transformer: Hierarchical Vision Transformer using Shifted Windows` -
441
+ https://arxiv.org/pdf/2103.14030
442
+
443
+ Args:
444
+ pretrain_img_size (int): Input image size for training the pretrained model,
445
+ used in absolute postion embedding. Default 224.
446
+ patch_size (int | tuple(int)): Patch size. Default: 4.
447
+ in_chans (int): Number of input image channels. Default: 3.
448
+ embed_dim (int): Number of linear projection output channels. Default: 96.
449
+ depths (tuple[int]): Depths of each Swin Transformer stage.
450
+ num_heads (tuple[int]): Number of attention head of each stage.
451
+ window_size (int): Window size. Default: 7.
452
+ mlp_ratio (float): Ratio of mlp hidden dim to embedding dim. Default: 4.
453
+ qkv_bias (bool): If True, add a learnable bias to query, key, value. Default: True
454
+ qk_scale (float): Override default qk scale of head_dim ** -0.5 if set.
455
+ drop_rate (float): Dropout rate.
456
+ attn_drop_rate (float): Attention dropout rate. Default: 0.
457
+ drop_path_rate (float): Stochastic depth rate. Default: 0.2.
458
+ norm_layer (nn.Module): Normalization layer. Default: nn.LayerNorm.
459
+ ape (bool): If True, add absolute position embedding to the patch embedding. Default: False.
460
+ patch_norm (bool): If True, add normalization after patch embedding. Default: True.
461
+ out_indices (Sequence[int]): Output from which stages.
462
+ frozen_stages (int): Stages to be frozen (stop grad and set eval mode).
463
+ -1 means not freezing any parameters.
464
+ use_checkpoint (bool): Whether to use checkpointing to save memory. Default: False.
465
+ """
466
+
467
+ def __init__(self,
468
+ pretrain_img_size=224,
469
+ patch_size=4,
470
+ in_chans=3,
471
+ embed_dim=96,
472
+ depths=[2, 2, 6, 2],
473
+ num_heads=[3, 6, 12, 24],
474
+ window_size=7,
475
+ mlp_ratio=4.,
476
+ qkv_bias=True,
477
+ qk_scale=None,
478
+ drop_rate=0.,
479
+ attn_drop_rate=0.,
480
+ drop_path_rate=0.2,
481
+ norm_layer=nn.LayerNorm,
482
+ ape=False,
483
+ patch_norm=True,
484
+ out_indices=(0, 1, 2, 3),
485
+ frozen_stages=-1,
486
+ use_checkpoint=False):
487
+ super().__init__()
488
+
489
+ self.pretrain_img_size = pretrain_img_size
490
+ self.num_layers = len(depths)
491
+ self.embed_dim = embed_dim
492
+ self.ape = ape
493
+ self.patch_norm = patch_norm
494
+ self.out_indices = out_indices
495
+ self.frozen_stages = frozen_stages
496
+
497
+ # split image into non-overlapping patches
498
+ self.patch_embed = PatchEmbed(
499
+ patch_size=patch_size, in_chans=in_chans, embed_dim=embed_dim,
500
+ norm_layer=norm_layer if self.patch_norm else None)
501
+
502
+ # absolute position embedding
503
+ if self.ape:
504
+ pretrain_img_size = to_2tuple(pretrain_img_size)
505
+ patch_size = to_2tuple(patch_size)
506
+ patches_resolution = [pretrain_img_size[0] // patch_size[0], pretrain_img_size[1] // patch_size[1]]
507
+
508
+ self.absolute_pos_embed = nn.Parameter(torch.zeros(1, embed_dim, patches_resolution[0], patches_resolution[1]))
509
+ trunc_normal_(self.absolute_pos_embed, std=.02)
510
+
511
+ self.pos_drop = nn.Dropout(p=drop_rate)
512
+
513
+ # stochastic depth
514
+ dpr = [x.item() for x in torch.linspace(0, drop_path_rate, sum(depths))] # stochastic depth decay rule
515
+
516
+ # build layers
517
+ self.layers = nn.ModuleList()
518
+ for i_layer in range(self.num_layers):
519
+ layer = BasicLayer(
520
+ dim=int(embed_dim * 2 ** i_layer),
521
+ depth=depths[i_layer],
522
+ num_heads=num_heads[i_layer],
523
+ window_size=window_size,
524
+ mlp_ratio=mlp_ratio,
525
+ qkv_bias=qkv_bias,
526
+ qk_scale=qk_scale,
527
+ drop=drop_rate,
528
+ attn_drop=attn_drop_rate,
529
+ drop_path=dpr[sum(depths[:i_layer]):sum(depths[:i_layer + 1])],
530
+ norm_layer=norm_layer,
531
+ downsample=PatchMerging if (i_layer < self.num_layers - 1) else None,
532
+ use_checkpoint=use_checkpoint)
533
+ self.layers.append(layer)
534
+
535
+ num_features = [int(embed_dim * 2 ** i) for i in range(self.num_layers)]
536
+ self.num_features = num_features
537
+
538
+ # add a norm layer for each output
539
+ for i_layer in out_indices:
540
+ layer = norm_layer(num_features[i_layer])
541
+ layer_name = f'norm{i_layer}'
542
+ self.add_module(layer_name, layer)
543
+
544
+ self._freeze_stages()
545
+
546
+ def _freeze_stages(self):
547
+ if self.frozen_stages >= 0:
548
+ self.patch_embed.eval()
549
+ for param in self.patch_embed.parameters():
550
+ param.requires_grad = False
551
+
552
+ if self.frozen_stages >= 1 and self.ape:
553
+ self.absolute_pos_embed.requires_grad = False
554
+
555
+ if self.frozen_stages >= 2:
556
+ self.pos_drop.eval()
557
+ for i in range(0, self.frozen_stages - 1):
558
+ m = self.layers[i]
559
+ m.eval()
560
+ for param in m.parameters():
561
+ param.requires_grad = False
562
+
563
+ def init_weights(self, pretrained=None):
564
+ """Initialize the weights in backbone.
565
+
566
+ Args:
567
+ pretrained (str, optional): Path to pre-trained weights.
568
+ Defaults to None.
569
+ """
570
+
571
+ def _init_weights(m):
572
+ if isinstance(m, nn.Linear):
573
+ trunc_normal_(m.weight, std=.02)
574
+ if isinstance(m, nn.Linear) and m.bias is not None:
575
+ nn.init.constant_(m.bias, 0)
576
+ elif isinstance(m, nn.LayerNorm):
577
+ nn.init.constant_(m.bias, 0)
578
+ nn.init.constant_(m.weight, 1.0)
579
+
580
+ if isinstance(pretrained, str):
581
+ self.apply(_init_weights)
582
+ # logger = get_root_logger()
583
+ load_checkpoint(self, pretrained, strict=False)
584
+ elif pretrained is None:
585
+ self.apply(_init_weights)
586
+ else:
587
+ raise TypeError('pretrained must be a str or None')
588
+
589
+ def forward(self, x):
590
+ """Forward function."""
591
+ x = self.patch_embed(x)
592
+
593
+ Wh, Ww = x.size(2), x.size(3)
594
+ if self.ape:
595
+ # interpolate the position embedding to the corresponding size
596
+ absolute_pos_embed = F.interpolate(self.absolute_pos_embed, size=(Wh, Ww), mode='bicubic')
597
+ x = (x + absolute_pos_embed).flatten(2).transpose(1, 2) # B Wh*Ww C
598
+ else:
599
+ x = x.flatten(2).transpose(1, 2)
600
+ x = self.pos_drop(x)
601
+
602
+ outs = []
603
+ for i in range(self.num_layers):
604
+ layer = self.layers[i]
605
+ x_out, H, W, x, Wh, Ww = layer(x, Wh, Ww)
606
+
607
+ if i in self.out_indices:
608
+ norm_layer = getattr(self, f'norm{i}')
609
+ x_out = norm_layer(x_out)
610
+
611
+ out = x_out.view(-1, H, W, self.num_features[i]).permute(0, 3, 1, 2).contiguous()
612
+ outs.append(out)
613
+
614
+ return tuple(outs)
615
+
616
+ def train(self, mode=True):
617
+ """Convert the model into training mode while keep layers freezed."""
618
+ super(SwinTransformer, self).train(mode)
619
+ self._freeze_stages()
external/WildCamera/WildCamera/newcrfs/uper_crf_head.py ADDED
@@ -0,0 +1,350 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import torch
2
+ import torch.nn as nn
3
+
4
+ from mmcv.cnn import ConvModule
5
+ from .newcrf_utils import resize
6
+
7
+
8
+ class PPM(nn.ModuleList):
9
+ """Pooling Pyramid Module used in PSPNet.
10
+
11
+ Args:
12
+ pool_scales (tuple[int]): Pooling scales used in Pooling Pyramid
13
+ Module.
14
+ in_channels (int): Input channels.
15
+ channels (int): Channels after modules, before conv_seg.
16
+ conv_cfg (dict|None): Config of conv layers.
17
+ norm_cfg (dict|None): Config of norm layers.
18
+ act_cfg (dict): Config of activation layers.
19
+ align_corners (bool): align_corners argument of F.interpolate.
20
+ """
21
+
22
+ def __init__(self, pool_scales, in_channels, channels, conv_cfg, norm_cfg,
23
+ act_cfg, align_corners):
24
+ super(PPM, self).__init__()
25
+ self.pool_scales = pool_scales
26
+ self.align_corners = align_corners
27
+ self.in_channels = in_channels
28
+ self.channels = channels
29
+ self.conv_cfg = conv_cfg
30
+ self.norm_cfg = norm_cfg
31
+ self.act_cfg = act_cfg
32
+ for pool_scale in pool_scales:
33
+ # == if batch size = 1, BN is not supported, change to GN
34
+ if pool_scale == 1: norm_cfg = dict(type='GN', requires_grad=True, num_groups=256)
35
+ self.append(
36
+ nn.Sequential(
37
+ nn.AdaptiveAvgPool2d(pool_scale),
38
+ ConvModule(
39
+ self.in_channels,
40
+ self.channels,
41
+ 1,
42
+ conv_cfg=self.conv_cfg,
43
+ norm_cfg=norm_cfg,
44
+ act_cfg=self.act_cfg)))
45
+
46
+ def forward(self, x):
47
+ """Forward function."""
48
+ ppm_outs = []
49
+ for ppm in self:
50
+ ppm_out = ppm(x)
51
+ upsampled_ppm_out = resize(
52
+ ppm_out,
53
+ size=x.size()[2:],
54
+ mode='bilinear',
55
+ align_corners=self.align_corners)
56
+ ppm_outs.append(upsampled_ppm_out)
57
+ return ppm_outs
58
+
59
+
60
+ class BaseDecodeHead(nn.Module):
61
+ """Base class for BaseDecodeHead.
62
+
63
+ Args:
64
+ in_channels (int|Sequence[int]): Input channels.
65
+ channels (int): Channels after modules, before conv_seg.
66
+ num_classes (int): Number of classes.
67
+ dropout_ratio (float): Ratio of dropout layer. Default: 0.1.
68
+ conv_cfg (dict|None): Config of conv layers. Default: None.
69
+ norm_cfg (dict|None): Config of norm layers. Default: None.
70
+ act_cfg (dict): Config of activation layers.
71
+ Default: dict(type='ReLU')
72
+ in_index (int|Sequence[int]): Input feature index. Default: -1
73
+ input_transform (str|None): Transformation type of input features.
74
+ Options: 'resize_concat', 'multiple_select', None.
75
+ 'resize_concat': Multiple feature maps will be resize to the
76
+ same size as first one and than concat together.
77
+ Usually used in FCN head of HRNet.
78
+ 'multiple_select': Multiple feature maps will be bundle into
79
+ a list and passed into decode head.
80
+ None: Only one select feature map is allowed.
81
+ Default: None.
82
+ loss_decode (dict): Config of decode loss.
83
+ Default: dict(type='CrossEntropyLoss').
84
+ ignore_index (int | None): The label index to be ignored. When using
85
+ masked BCE loss, ignore_index should be set to None. Default: 255
86
+ sampler (dict|None): The config of segmentation map sampler.
87
+ Default: None.
88
+ align_corners (bool): align_corners argument of F.interpolate.
89
+ Default: False.
90
+ """
91
+
92
+ def __init__(self,
93
+ in_channels,
94
+ channels,
95
+ *,
96
+ num_classes,
97
+ dropout_ratio=0.1,
98
+ conv_cfg=None,
99
+ norm_cfg=None,
100
+ act_cfg=dict(type='ReLU'),
101
+ in_index=-1,
102
+ input_transform=None,
103
+ loss_decode=None,
104
+ ignore_index=255,
105
+ sampler=None,
106
+ align_corners=False):
107
+ super(BaseDecodeHead, self).__init__()
108
+ self._init_inputs(in_channels, in_index, input_transform)
109
+ self.channels = channels
110
+ self.num_classes = num_classes
111
+ self.dropout_ratio = dropout_ratio
112
+ self.conv_cfg = conv_cfg
113
+ self.norm_cfg = norm_cfg
114
+ self.act_cfg = act_cfg
115
+ self.in_index = in_index
116
+ self.ignore_index = ignore_index
117
+ self.align_corners = align_corners
118
+ if dropout_ratio > 0:
119
+ self.dropout = nn.Dropout2d(dropout_ratio)
120
+ else:
121
+ self.dropout = None
122
+ self.fp16_enabled = False
123
+
124
+ def extra_repr(self):
125
+ """Extra repr."""
126
+ s = f'input_transform={self.input_transform}, ' \
127
+ f'ignore_index={self.ignore_index}, ' \
128
+ f'align_corners={self.align_corners}'
129
+ return s
130
+
131
+ def _init_inputs(self, in_channels, in_index, input_transform):
132
+ """Check and initialize input transforms.
133
+
134
+ The in_channels, in_index and input_transform must match.
135
+ Specifically, when input_transform is None, only single feature map
136
+ will be selected. So in_channels and in_index must be of type int.
137
+ When input_transform
138
+
139
+ Args:
140
+ in_channels (int|Sequence[int]): Input channels.
141
+ in_index (int|Sequence[int]): Input feature index.
142
+ input_transform (str|None): Transformation type of input features.
143
+ Options: 'resize_concat', 'multiple_select', None.
144
+ 'resize_concat': Multiple feature maps will be resize to the
145
+ same size as first one and than concat together.
146
+ Usually used in FCN head of HRNet.
147
+ 'multiple_select': Multiple feature maps will be bundle into
148
+ a list and passed into decode head.
149
+ None: Only one select feature map is allowed.
150
+ """
151
+
152
+ if input_transform is not None:
153
+ assert input_transform in ['resize_concat', 'multiple_select']
154
+ self.input_transform = input_transform
155
+ self.in_index = in_index
156
+ if input_transform is not None:
157
+ assert isinstance(in_channels, (list, tuple))
158
+ assert isinstance(in_index, (list, tuple))
159
+ assert len(in_channels) == len(in_index)
160
+ if input_transform == 'resize_concat':
161
+ self.in_channels = sum(in_channels)
162
+ else:
163
+ self.in_channels = in_channels
164
+ else:
165
+ assert isinstance(in_channels, int)
166
+ assert isinstance(in_index, int)
167
+ self.in_channels = in_channels
168
+
169
+ def init_weights(self):
170
+ """Initialize weights of classification layer."""
171
+ # normal_init(self.conv_seg, mean=0, std=0.01)
172
+ # normal_init(self.conv1, mean=0, std=0.01)
173
+
174
+ def _transform_inputs(self, inputs):
175
+ """Transform inputs for decoder.
176
+
177
+ Args:
178
+ inputs (list[Tensor]): List of multi-level img features.
179
+
180
+ Returns:
181
+ Tensor: The transformed inputs
182
+ """
183
+
184
+ if self.input_transform == 'resize_concat':
185
+ inputs = [inputs[i] for i in self.in_index]
186
+ upsampled_inputs = [
187
+ resize(
188
+ input=x,
189
+ size=inputs[0].shape[2:],
190
+ mode='bilinear',
191
+ align_corners=self.align_corners) for x in inputs
192
+ ]
193
+ inputs = torch.cat(upsampled_inputs, dim=1)
194
+ elif self.input_transform == 'multiple_select':
195
+ inputs = [inputs[i] for i in self.in_index]
196
+ else:
197
+ inputs = inputs[self.in_index]
198
+
199
+ return inputs
200
+
201
+ def forward(self, inputs):
202
+ """Placeholder of forward function."""
203
+ pass
204
+
205
+ def forward_train(self, inputs, img_metas, gt_semantic_seg, train_cfg):
206
+ """Forward function for training.
207
+ Args:
208
+ inputs (list[Tensor]): List of multi-level img features.
209
+ img_metas (list[dict]): List of image info dict where each dict
210
+ has: 'img_shape', 'scale_factor', 'flip', and may also contain
211
+ 'filename', 'ori_shape', 'pad_shape', and 'img_norm_cfg'.
212
+ For details on the values of these keys see
213
+ `mmseg/datasets/pipelines/formatting.py:Collect`.
214
+ gt_semantic_seg (Tensor): Semantic segmentation masks
215
+ used if the architecture supports semantic segmentation task.
216
+ train_cfg (dict): The training config.
217
+
218
+ Returns:
219
+ dict[str, Tensor]: a dictionary of loss components
220
+ """
221
+ seg_logits = self.forward(inputs)
222
+ losses = self.losses(seg_logits, gt_semantic_seg)
223
+ return losses
224
+
225
+ def forward_test(self, inputs, img_metas, test_cfg):
226
+ """Forward function for testing.
227
+
228
+ Args:
229
+ inputs (list[Tensor]): List of multi-level img features.
230
+ img_metas (list[dict]): List of image info dict where each dict
231
+ has: 'img_shape', 'scale_factor', 'flip', and may also contain
232
+ 'filename', 'ori_shape', 'pad_shape', and 'img_norm_cfg'.
233
+ For details on the values of these keys see
234
+ `mmseg/datasets/pipelines/formatting.py:Collect`.
235
+ test_cfg (dict): The testing config.
236
+
237
+ Returns:
238
+ Tensor: Output segmentation map.
239
+ """
240
+ return self.forward(inputs)
241
+
242
+
243
+ class UPerHead(BaseDecodeHead):
244
+ def __init__(self, pool_scales=(1, 2, 3, 6), **kwargs):
245
+ super(UPerHead, self).__init__(
246
+ input_transform='multiple_select', **kwargs)
247
+ # FPN Module
248
+ self.lateral_convs = nn.ModuleList()
249
+ self.fpn_convs = nn.ModuleList()
250
+ for in_channels in self.in_channels: # skip the top layer
251
+ l_conv = ConvModule(
252
+ in_channels,
253
+ self.channels,
254
+ 1,
255
+ conv_cfg=self.conv_cfg,
256
+ norm_cfg=self.norm_cfg,
257
+ act_cfg=self.act_cfg,
258
+ inplace=True)
259
+ fpn_conv = ConvModule(
260
+ self.channels,
261
+ self.channels,
262
+ 3,
263
+ padding=1,
264
+ conv_cfg=self.conv_cfg,
265
+ norm_cfg=self.norm_cfg,
266
+ act_cfg=self.act_cfg,
267
+ inplace=True)
268
+ self.lateral_convs.append(l_conv)
269
+ self.fpn_convs.append(fpn_conv)
270
+
271
+ def forward(self, inputs):
272
+ """Forward function."""
273
+
274
+ inputs = self._transform_inputs(inputs)
275
+
276
+ # build laterals
277
+ laterals = [
278
+ lateral_conv(inputs[i])
279
+ for i, lateral_conv in enumerate(self.lateral_convs)
280
+ ]
281
+
282
+ # laterals.append(self.psp_forward(inputs))
283
+
284
+ # build top-down path
285
+ used_backbone_levels = len(laterals)
286
+ for i in range(used_backbone_levels - 1, 0, -1):
287
+ prev_shape = laterals[i - 1].shape[2:]
288
+ laterals[i - 1] += resize(
289
+ laterals[i],
290
+ size=prev_shape,
291
+ mode='bilinear',
292
+ align_corners=self.align_corners)
293
+
294
+ # build outputs
295
+ fpn_outs = [
296
+ self.fpn_convs[i](laterals[i])
297
+ for i in range(used_backbone_levels - 1)
298
+ ]
299
+ # append psp feature
300
+ fpn_outs.append(laterals[-1])
301
+
302
+ return fpn_outs[0]
303
+
304
+ class PSP(BaseDecodeHead):
305
+ """Unified Perceptual Parsing for Scene Understanding.
306
+
307
+ This head is the implementation of `UPerNet
308
+ <https://arxiv.org/abs/1807.10221>`_.
309
+
310
+ Args:
311
+ pool_scales (tuple[int]): Pooling scales used in Pooling Pyramid
312
+ Module applied on the last feature. Default: (1, 2, 3, 6).
313
+ """
314
+
315
+ def __init__(self, pool_scales=(1, 2, 3, 6), **kwargs):
316
+ super(PSP, self).__init__(
317
+ input_transform='multiple_select', **kwargs)
318
+ # PSP Module
319
+ self.psp_modules = PPM(
320
+ pool_scales,
321
+ self.in_channels[-1],
322
+ self.channels,
323
+ conv_cfg=self.conv_cfg,
324
+ norm_cfg=self.norm_cfg,
325
+ act_cfg=self.act_cfg,
326
+ align_corners=self.align_corners)
327
+ self.bottleneck = ConvModule(
328
+ self.in_channels[-1] + len(pool_scales) * self.channels,
329
+ self.channels,
330
+ 3,
331
+ padding=1,
332
+ conv_cfg=self.conv_cfg,
333
+ norm_cfg=self.norm_cfg,
334
+ act_cfg=self.act_cfg)
335
+
336
+ def psp_forward(self, inputs):
337
+ """Forward function of PSP module."""
338
+ x = inputs[-1]
339
+ psp_outs = [x]
340
+ psp_outs.extend(self.psp_modules(x))
341
+ psp_outs = torch.cat(psp_outs, dim=1)
342
+ output = self.bottleneck(psp_outs)
343
+
344
+ return output
345
+
346
+ def forward(self, inputs):
347
+ """Forward function."""
348
+ inputs = self._transform_inputs(inputs)
349
+
350
+ return self.psp_forward(inputs)
external/WildCamera/WildCamera/train/train_calibrator.py ADDED
@@ -0,0 +1,297 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from __future__ import print_function, division
2
+ import os, sys, inspect, time, copy, warnings
3
+ project_root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe())))))
4
+ sys.path.insert(0, project_root)
5
+ warnings.filterwarnings('ignore')
6
+
7
+ import torch
8
+ import torch.backends.cudnn as cudnn
9
+ import torch.distributed as dist
10
+ import torch.multiprocessing as mp
11
+ from torch.utils.tensorboard import SummaryWriter
12
+ from torch.utils.data import ConcatDataset
13
+ from torchvision import transforms
14
+
15
+ import argparse
16
+ import numpy as np
17
+ from loguru import logger
18
+ from einops import rearrange
19
+ from pprint import pprint
20
+
21
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
22
+ from WildCamera.datasets.IncdDataset import IncdDataset
23
+ from tools.tools import to_cuda, intrinsic2incidence, resample_rgb, DistributedSamplerNoEvenlyDivisible, IncidenceLoss
24
+ from tools.visualization import tensor2disp, tensor2rgb
25
+
26
+ parser = argparse.ArgumentParser(description='NeWCRFs PyTorch implementation.', fromfile_prefix_chars='@')
27
+
28
+ parser.add_argument('--experiment_name', type=str, help='experiment name', default='multi-modality newcrfs')
29
+ parser.add_argument('--experiment_set', type=str, choices=['gsv', 'in_the_wild'], required=True)
30
+ parser.add_argument('--saving_location', type=str, help='saving location', default=None)
31
+ parser.add_argument('--encoder', type=str, help='type of encoder, base07, large07', default='large07')
32
+ parser.add_argument('--pretrain', type=str, help='path of pretrained encoder', default='model_zoo/swin_transformer/swin_large_patch4_window7_224_22k.pth')
33
+ parser.add_argument('--load_ckpt', type=str, help='path of ckpt', default=None)
34
+ parser.add_argument('--evaluation_only', action="store_true")
35
+ parser.add_argument('--l1_th', type=int, help='RANSAC threshold', default=0.02)
36
+
37
+ # Dataset
38
+ parser.add_argument('--data_path', type=str, help='path to the data', default='data/MonoCalib')
39
+ parser.add_argument('--input_height', type=int, help='input height', default=480)
40
+ parser.add_argument('--input_width', type=int, help='input width', default=640)
41
+
42
+ # Training
43
+ parser.add_argument('--batch_size', type=int, help='batch size', default=16)
44
+ parser.add_argument('--loss', type=str, default='cosine', help='You can also choees l1 loss')
45
+ parser.add_argument('--steps_per_epoch', type=int, help='frequency for evaluation', default=1000)
46
+ parser.add_argument('--termination_epoch', type=int, help='epoch to stop training', default=25)
47
+
48
+ # Training Misc
49
+ parser.add_argument('--weight_decay', type=float, help='weight decay factor for optimization', default=1e-2)
50
+ parser.add_argument('--adam_eps', type=float, help='epsilon in Adam optimizer', default=1e-6)
51
+ parser.add_argument('--train_workers', type=int, help='dataloader workers', default=32)
52
+ parser.add_argument('--eval_workers', type=int, help='dataloader workers', default=2)
53
+ parser.add_argument('--learning_rate', type=float, help='initial learning rate', default=1e-5)
54
+ parser.add_argument('--end_learning_rate', type=float, help='end learning rate', default=-1)
55
+
56
+ # Augmentation
57
+ parser.add_argument('--dataset_favour_long', type=float, default=0.1, help='whether in training will see more samples from large dataset')
58
+ parser.add_argument('--augscale', type=float, default=2.0, help='The scale of Augmentation')
59
+ parser.add_argument('--no_change_prob', type=float, default=0.1, help='The probability of seeing original image')
60
+ parser.add_argument('--coloraugmentation', action="store_true")
61
+ parser.add_argument('--coloraugmentation_scale', type=float, default=0.0)
62
+
63
+ # Multi-gpu training
64
+ parser.add_argument('--gpu', type=int, help='GPU id to use.', default=None)
65
+ parser.add_argument('--dist_url', type=str, help='url used to set up distributed training', default='tcp://127.0.0.1:1235')
66
+ parser.add_argument('--dist_backend', type=str, help='distributed backend', default='nccl')
67
+
68
+ def main_worker(gpu, ngpus_per_node, args):
69
+ args.gpu = gpu
70
+ group = dist.init_process_group(backend=args.dist_backend, init_method=args.dist_url, world_size=args.world_size, rank=args.gpu)
71
+
72
+ # NeWCRFs model
73
+ model = NEWCRFIF(version=args.encoder, pretrained=args.pretrain)
74
+ model.train()
75
+
76
+ if args.load_ckpt is not None:
77
+ model.load_state_dict(torch.load(args.load_ckpt, map_location="cpu"), strict=True)
78
+ model.eval()
79
+ logger.info("Load Model from %s" % args.load_ckpt)
80
+
81
+ if args.distributed:
82
+ if args.gpu is not None:
83
+ torch.cuda.set_device(args.gpu)
84
+ model.cuda(args.gpu)
85
+ args.batch_size = int(args.batch_size / ngpus_per_node)
86
+ args.train_workers = int(args.train_workers / ngpus_per_node)
87
+ model = torch.nn.parallel.DistributedDataParallel(model, device_ids=[args.gpu], find_unused_parameters=False)
88
+ else:
89
+ model.cuda()
90
+ model = torch.nn.parallel.DistributedDataParallel(model, find_unused_parameters=False)
91
+ else:
92
+ model = torch.nn.DataParallel(model)
93
+ model.cuda()
94
+
95
+ if args.distributed:
96
+ print("== Model Initialized on GPU: {}".format(args.gpu))
97
+ else:
98
+ print("== Model Initialized")
99
+
100
+ # Training parameters
101
+ optimizer = torch.optim.Adam([{'params': model.module.parameters()}], lr=args.learning_rate)
102
+
103
+ cudnn.benchmark = True
104
+
105
+ incidence_criterion = IncidenceLoss(loss=args.loss)
106
+
107
+ end_learning_rate = args.end_learning_rate if args.end_learning_rate != -1 else 0.1 * args.learning_rate
108
+
109
+ steps_per_epoch = args.steps_per_epoch
110
+ num_total_steps = 250000
111
+ epoch = int(num_total_steps / steps_per_epoch)
112
+
113
+ inv_normalize = transforms.Normalize(
114
+ mean=[-0.485 / 0.229, -0.456 / 0.224, -0.406 / 0.225],
115
+ std=[1 / 0.229, 1 / 0.224, 1 / 0.225]
116
+ )
117
+
118
+ if args.gpu == 0:
119
+ checkpoint_dir = os.path.join(args.saving_location, 'model_zoo', args.experiment_name)
120
+ writer = SummaryWriter(checkpoint_dir, flush_secs=30)
121
+ logger.add(os.path.join(checkpoint_dir, "{}.log".format(args.experiment_name)))
122
+ else:
123
+ writer = None
124
+ logger.remove()
125
+ logger.add(sys.stderr, level="ERROR")
126
+
127
+ if args.experiment_set == 'gsv':
128
+ from WildCamera.evaluation.evaluate_fov import EvaluateFov
129
+ evaluator = EvaluateFov()
130
+ elif args.experiment_set == 'in_the_wild':
131
+ from WildCamera.evaluation.evaluate_intrinsic import EvaluateIntrinsic
132
+ evaluator = EvaluateIntrinsic()
133
+ else:
134
+ raise NotImplementedError()
135
+
136
+ if args.evaluation_only:
137
+ evaluator.evaluate(
138
+ model,
139
+ args,
140
+ steps=0,
141
+ writer=writer,
142
+ group=group,
143
+ wtassumption=False,
144
+ )
145
+ evaluator.evaluate(
146
+ model,
147
+ args,
148
+ steps=0,
149
+ writer=writer,
150
+ group=group,
151
+ wtassumption=True,
152
+ )
153
+ return
154
+
155
+ for n_0 in range(epoch):
156
+
157
+ if n_0 > args.termination_epoch:
158
+ break
159
+
160
+ incdataset = IncdDataset(
161
+ data_root=args.data_path,
162
+ datasets_included=args.datasets_train,
163
+ ht=args.input_height,
164
+ wt=args.input_width,
165
+ augmentation=True,
166
+ split='train',
167
+ shuffleseed=int(n_0 + 1),
168
+ dataset_favour_long=args.dataset_favour_long,
169
+ augscale=args.augscale,
170
+ no_change_prob=args.no_change_prob,
171
+ coloraugmentation=args.coloraugmentation,
172
+ coloraugmentation_scale=args.coloraugmentation_scale
173
+ )
174
+ sampler = DistributedSamplerNoEvenlyDivisible(incdataset, shuffle=True)
175
+ dataloader = torch.utils.data.DataLoader(
176
+ incdataset,
177
+ sampler=sampler,
178
+ batch_size=int(args.batch_size),
179
+ num_workers=int(args.train_workers)
180
+ )
181
+ assert len(dataloader) > steps_per_epoch
182
+ dataloader = iter(dataloader)
183
+ sampler.set_epoch(n_0)
184
+ for global_step in range(int(n_0 * steps_per_epoch), int((n_0 + 1) * steps_per_epoch)):
185
+ sample_batched = next(dataloader)
186
+ sample_batched = to_cuda(sample_batched)
187
+
188
+ rgb, K = sample_batched['rgb'], sample_batched['K']
189
+
190
+ model.train()
191
+ incidence = model(rgb)
192
+ optimizer.zero_grad()
193
+
194
+ # Loss For Normal
195
+ loss = incidence_criterion(incidence, K)
196
+ loss.backward()
197
+
198
+ for param_group in optimizer.param_groups:
199
+ current_lr = (args.learning_rate - end_learning_rate) * (1 - global_step / num_total_steps) ** 0.9 + end_learning_rate
200
+ param_group['lr'] = current_lr
201
+
202
+ optimizer.step()
203
+
204
+ if np.mod(global_step, 1000) == 0 and args.gpu == 0:
205
+ b = 1
206
+ _, _, h, w = rgb.shape
207
+
208
+ rgb = inv_normalize(rgb)
209
+ vls1 = tensor2rgb(rgb, viewind=0)
210
+
211
+ device = rgb.device
212
+ incidence_gt = intrinsic2incidence(K[0:1], b, h, w, device)
213
+ incidence_gt = rearrange(incidence_gt.squeeze(dim=4), 'b h w d -> b d h w')
214
+ vls3 = tensor2rgb((incidence + 1) / 2, viewind=0)
215
+ vls4 = tensor2rgb((incidence_gt + 1) / 2, viewind=0)
216
+
217
+ vls = np.concatenate([np.array(vls1), np.array(vls3), np.array(vls4)], axis=0)
218
+ writer.add_image('visualization', (torch.from_numpy(vls).float() / 255).permute([2, 0, 1]), global_step)
219
+
220
+ if writer is not None and args.gpu == 0:
221
+ writer.add_scalar('loss/loss_incidence', loss.item(), global_step)
222
+ if np.mod(global_step, 500) == 0:
223
+ logger.info('Step %d, Epoch %d, loss %.3f' % (global_step, n_0, loss.item()))
224
+
225
+ evaluator.evaluate(
226
+ model,
227
+ args,
228
+ steps=global_step,
229
+ writer=writer,
230
+ group=group,
231
+ wtassumption=False,
232
+ )
233
+
234
+ evaluator.evaluate(
235
+ model,
236
+ args,
237
+ steps=global_step,
238
+ writer=writer,
239
+ group=group,
240
+ wtassumption=True,
241
+ )
242
+
243
+ def main():
244
+ args = parser.parse_args()
245
+ torch.cuda.empty_cache()
246
+ args.world_size = torch.cuda.device_count()
247
+ args.distributed = True
248
+ args.dist_url = 'tcp://127.0.0.1:' + str(np.random.randint(2000, 3000, 1).item())
249
+
250
+ if args.saving_location is None:
251
+ args.saving_location = project_root
252
+
253
+ if args.experiment_set == 'gsv':
254
+ args.datasets_train = [
255
+ 'GSV'
256
+ ]
257
+ args.datasets_eval = [
258
+ 'GSV'
259
+ ]
260
+ elif args.experiment_set == 'in_the_wild':
261
+ args.datasets_train = [
262
+ 'Nuscenes',
263
+ 'KITTI',
264
+ 'Cityscapes',
265
+ 'NYUv2',
266
+ 'ARKitScenes',
267
+ 'MegaDepth',
268
+ 'SUN3D',
269
+ 'MVImgNet',
270
+ 'Objectron'
271
+ ]
272
+ args.datasets_eval = [
273
+ 'Nuscenes',
274
+ 'KITTI',
275
+ 'Cityscapes',
276
+ 'NYUv2',
277
+ 'ARKitScenes',
278
+ 'MegaDepth',
279
+ 'SUN3D',
280
+ 'MVImgNet',
281
+ 'Objectron',
282
+ 'Waymo',
283
+ 'BIWIRGBDID',
284
+ 'RGBD',
285
+ 'ScanNet',
286
+ 'MVS',
287
+ 'Scenes11'
288
+ ]
289
+
290
+ else:
291
+ raise NotImplementedError()
292
+
293
+ pprint(vars(args))
294
+ mp.spawn(main_worker, nprocs=args.world_size, args=(args.world_size, args))
295
+
296
+ if __name__ == '__main__':
297
+ main()
external/WildCamera/__pycache__/hubconf.cpython-310.pyc ADDED
Binary file (720 Bytes). View file
 
external/WildCamera/asset/download_demo_images.sh ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ cd asset
2
+ wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/asset/dollyzoom.tar
3
+ tar -xvf dollyzoom.tar
4
+ rm dollyzoom.tar
5
+ wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/asset/images-from-github-wt-intrinsic.tar
6
+ tar -xvf images-from-github-wt-intrinsic.tar
7
+ rm images-from-github-wt-intrinsic.tar
8
+ cd ..
external/WildCamera/asset/download_wildcamera_checkpoint.sh ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ cd model_zoo
2
+ mkdir Release
3
+ cd Release
4
+ wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/wild_camera_all.pth?download=true
5
+ mv wild_camera_all.pth?download=true wild_camera_all.pth
6
+ wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/wild_camera_gsv.pth?download=true
7
+ mv wild_camera_gsv.pth?download=true wild_camera_gsv.pth
8
+ cd ..
9
+ mkdir swin_transformer
10
+ cd swin_transformer
11
+ wget https://huggingface.co/datasets/Shengjie/WildCamera/resolve/main/checkpoint/swin_large_patch4_window7_224_22k.pth?download=true
12
+ mv swin_large_patch4_window7_224_22k.pth?download=true swin_large_patch4_window7_224_22k.pth
13
+ cd ..
external/WildCamera/asset/download_wildcamera_dataset.sh ADDED
The diff for this file is too large to render. See raw diff
 
external/WildCamera/demo/demo_dollyzoom.py ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, copy, tqdm, glob, natsort, torch
2
+ import numpy as np
3
+ project_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
4
+
5
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
6
+ from PIL import Image, ImageDraw, ImageFont
7
+
8
+ def draw_focalbar(rgbaug, fest, minfocal, maxfocal):
9
+ tmp = copy.deepcopy(rgbaug)
10
+
11
+ font_size = 28
12
+ font = ImageFont.truetype("arial.ttf", size=font_size)
13
+
14
+ rgbaug = copy.deepcopy(tmp)
15
+ w, h = rgbaug.size
16
+ rgbaug.putalpha(255)
17
+
18
+ paddingr = 0.1
19
+ barsty = h * paddingr
20
+ baredy = h * (1 - paddingr)
21
+ barx = w * 0.9
22
+ horizonbarnum = 7 * 5
23
+ horizonbarlen = 0.01
24
+
25
+ white = Image.new('RGBA', rgbaug.size, (255, 255, 255, 0))
26
+ draw = ImageDraw.Draw(white)
27
+ draw.rectangle(
28
+ [
29
+ (w * 0.9 - w * horizonbarlen - 15 / 640 * w, h * paddingr - 15 / 480 * h),
30
+ w * 0.9 + w * horizonbarlen + 35 / 640 * w, h * (1 - paddingr) + 15 / 480 * h
31
+ ],
32
+ fill=(255, 255, 255, 128)
33
+ )
34
+ rgbaug = Image.alpha_composite(rgbaug, white)
35
+
36
+ draw = ImageDraw.Draw(rgbaug)
37
+ draw.line((barx, barsty, barx, baredy), fill=(0, 0, 0, 255), width=5)
38
+
39
+ for i in range(horizonbarnum + 1):
40
+ r = i / horizonbarnum
41
+ bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
42
+ barstx = w * 0.9 - w * horizonbarlen
43
+ baredx = w * 0.9 + w * horizonbarlen
44
+ draw.line((barstx, bary, baredx, bary), fill=(0, 0, 0, 255), width=int(3))
45
+
46
+ horizonbarnum = 7
47
+ horizonbarlen = 0.02
48
+
49
+ for i in range(horizonbarnum + 1):
50
+ r = i / horizonbarnum
51
+ bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
52
+ barstx = w * 0.9 - w * horizonbarlen
53
+ baredx = w * 0.9 + w * horizonbarlen
54
+ draw.line((barstx, bary, baredx, bary), fill=(0, 0, 0, 255), width=3)
55
+
56
+ textpadding = 30 / 640 * w
57
+ draw.text((barstx + textpadding, bary - 5 / 480 * h), str(int(maxfocal * r + minfocal * (1 - r))), fill=(0, 0, 0, 255), font=font)
58
+
59
+ r = (fest - minfocal) / (maxfocal - minfocal)
60
+ bary = h * paddingr * r + h * (1 - paddingr) * (1 - r)
61
+ barstx = w * 0.9 - w * horizonbarlen
62
+ baredx = w * 0.9 + w * horizonbarlen
63
+ draw.line((barstx, bary, baredx, bary), fill=(255, 165, 0, 255), width=5)
64
+ return rgbaug
65
+
66
+ def smooth_width(fests_precomputed, idx, width=3):
67
+ # Smooth as an average of Three Frames
68
+ stidx = idx - int((width - 1) / 2)
69
+ edidx = idx + int((width - 1) / 2)
70
+
71
+ if stidx < 0:
72
+ stidx = 0
73
+
74
+ if edidx >= len(fests_precomputed):
75
+ edidx = len(fests_precomputed) - 1
76
+
77
+ return np.mean(fests_precomputed[stidx:edidx+1])
78
+
79
+
80
+ @torch.no_grad()
81
+ def draw_dollyzoom(model):
82
+ """ In Validation, random sample N Images to mimic a real test set situation """
83
+ model.eval()
84
+
85
+ dollyzoom_folder = os.path.join(project_dir, 'asset', 'dollyzoom')
86
+ output_dollyzoom = os.path.join(project_dir, 'output', 'dollyzoom')
87
+ dollyzoomvideos = [
88
+ 'dz1',
89
+ 'dz2',
90
+ 'dz3'
91
+ ]
92
+
93
+ for dollyzoomvideo in dollyzoomvideos:
94
+ dollyzoomimg_folder = os.path.join(dollyzoom_folder, dollyzoomvideo)
95
+ jpgs = glob.glob(os.path.join(dollyzoomimg_folder, '*.jpg'))
96
+ jpgs = natsort.natsorted(jpgs)
97
+
98
+ fests = list()
99
+ for jpg in tqdm.tqdm(jpgs):
100
+ intrinsic, _ = model.inference(Image.open(jpg), wtassumption=False)
101
+ fest = intrinsic[0, 0]
102
+ fests.append(fest)
103
+
104
+ fests = np.array(fests)
105
+ minfocal = fests.min()
106
+ maxfocal = fests.max()
107
+
108
+ focalpadding = 0.1 * minfocal
109
+
110
+ jpgs_focalbar = list()
111
+ for idx, jpg in enumerate(jpgs):
112
+ fest = smooth_width(fests, idx, width=3)
113
+ jpgs_focalbar.append(draw_focalbar(Image.open(jpg), fest, minfocal=int(minfocal - focalpadding), maxfocal=int(maxfocal + focalpadding)))
114
+
115
+ output_folder = os.path.join(output_dollyzoom, dollyzoomvideo)
116
+ os.makedirs(output_folder, exist_ok=True)
117
+ for idx, jpgimgage in enumerate(jpgs_focalbar):
118
+ jpg_path = os.path.join(output_folder, '{}.png'.format(str(idx)))
119
+ jpgimgage.save(jpg_path)
120
+
121
+ return
122
+
123
+ if __name__ == '__main__':
124
+ # NeWCRFs model
125
+ model = NEWCRFIF(version='large07', pretrained=None)
126
+ model.eval()
127
+ model.cuda()
128
+
129
+ script_dir = os.path.dirname(os.path.realpath(__file__))
130
+ ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
131
+ model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
132
+
133
+ draw_dollyzoom(model)
external/WildCamera/demo/demo_inference.py ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, torch
2
+ project_dir = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
3
+
4
+ from PIL import Image
5
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
6
+
7
+
8
+ if __name__ == '__main__':
9
+ # NeWCRFs model
10
+ model = NEWCRFIF(version='large07', pretrained=None)
11
+ model.eval()
12
+ model.cuda()
13
+
14
+ script_dir = os.path.dirname(os.path.realpath(__file__))
15
+ ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
16
+ model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
17
+
18
+ images_folder = os.path.join(project_dir, 'asset', 'images-from-github-wt-intrinsic')
19
+ info_path = os.path.join(images_folder, 'intrinsic_gt.txt')
20
+ with open(info_path) as file:
21
+ infos = [line.rstrip() for line in file]
22
+
23
+ for idx, info in enumerate(infos):
24
+ imgname, focalgt, source = info.split(' ')
25
+
26
+ images_path = os.path.join(images_folder, imgname)
27
+ intrinsic, _ = model.inference(Image.open(images_path), wtassumption=False)
28
+ focal = intrinsic[0, 0].item()
29
+
30
+ print("Image Name: %s, Est Focal %.1f, Gt Focal %.1f, Source - %s" % (imgname, focal, float(focalgt), source))
external/WildCamera/demo/demo_restoration.py ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os, sys, inspect
2
+ project_root = os.path.dirname(os.path.dirname(os.path.abspath(inspect.getfile(inspect.currentframe()))))
3
+ sys.path.insert(0, project_root)
4
+
5
+ import PIL.Image as Image
6
+ import torch
7
+ from WildCamera.newcrfs.newcrf_incidencefield import NEWCRFIF
8
+
9
+ @torch.no_grad()
10
+ def main():
11
+ model = NEWCRFIF(version='large07', pretrained=None)
12
+ model.eval()
13
+ model.cuda()
14
+
15
+ script_dir = os.path.dirname(os.path.realpath(__file__))
16
+ ckpt_path = os.path.join(os.path.dirname(script_dir), 'model_zoo/Release', 'wild_camera_all.pth')
17
+ model.load_state_dict(torch.load(ckpt_path, map_location="cpu"), strict=True)
18
+
19
+ img_folder = os.path.join(project_root, 'asset', 'image_restoration')
20
+ img_names = [
21
+ 'lenna_distorted_cropright.jpg',
22
+ 'lenna_distorted_cropleft.jpg',
23
+ ]
24
+ for img_name in img_names:
25
+ img_path = os.path.join(img_folder, img_name)
26
+ img = Image.open(img_path)
27
+
28
+ export_root = os.path.join(img_folder, '{}_restored.jpg'.format(img_name.split('.')[0]))
29
+ intrinsic, _ = model.inference(img, wtassumption=False)
30
+ model.restore_image(img, intrinsic).save(export_root)
31
+
32
+ if __name__ == '__main__':
33
+ main()
external/WildCamera/splits/arkitscenes_test.txt ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ARKitScenes_test Validation_41069021_401.371_00005806
2
+ ARKitScenes_test Validation_41069021_412.367_00006466
3
+ ARKitScenes_test Validation_41069021_417.365_00006766
4
+ ARKitScenes_test Validation_41069021_421.363_00007006
5
+ ARKitScenes_test Validation_41069021_423.362_00007126
6
+ ARKitScenes_test Validation_41069021_435.357_00007846
7
+ ARKitScenes_test Validation_41069021_450.368_00008747
8
+ ARKitScenes_test Validation_41069025_624.863_00002137
9
+ ARKitScenes_test Validation_41069025_648.870_00003578
10
+ ARKitScenes_test Validation_41069025_731.869_00008560
11
+ ARKitScenes_test Validation_41069042_3044.239_00000085
12
+ ARKitScenes_test Validation_41069042_3045.222_00000144
13
+ ARKitScenes_test Validation_41069042_3063.231_00001225
14
+ ARKitScenes_test Validation_41069042_3070.228_00001645
15
+ ARKitScenes_test Validation_41069042_3073.227_00001825
16
+ ARKitScenes_test Validation_41069043_2820.631_00001297
17
+ ARKitScenes_test Validation_41069043_2826.629_00001657
18
+ ARKitScenes_test Validation_41069043_2830.627_00001897
19
+ ARKitScenes_test Validation_41069043_2892.635_00005619
20
+ ARKitScenes_test Validation_41069046_2929.336_00001117
21
+ ARKitScenes_test Validation_41069046_2988.328_00004658
22
+ ARKitScenes_test Validation_41069046_2989.328_00004718
23
+ ARKitScenes_test Validation_41069046_2997.325_00005198
24
+ ARKitScenes_test Validation_41069046_3024.330_00006819
25
+ ARKitScenes_test Validation_41069051_5085.217_00000211
26
+ ARKitScenes_test Validation_41142278_4003.081_00000097
27
+ ARKitScenes_test Validation_41142278_4070.070_00004118
28
+ ARKitScenes_test Validation_41142280_3837.982_00000937
29
+ ARKitScenes_test Validation_41142281_3964.380_00002798
30
+ ARKitScenes_test Validation_42444946_223634.869_00001657
31
+ ARKitScenes_test Validation_42444946_223654.878_00002858
32
+ ARKitScenes_test Validation_42444946_223693.879_00005199
33
+ ARKitScenes_test Validation_42444950_223834.571_00002198
34
+ ARKitScenes_test Validation_42444950_223862.577_00003879
35
+ ARKitScenes_test Validation_42444966_40376.779_00002863
36
+ ARKitScenes_test Validation_42444966_40381.794_00003164
37
+ ARKitScenes_test Validation_42444966_40382.793_00003224
38
+ ARKitScenes_test Validation_42444966_40391.790_00003764
39
+ ARKitScenes_test Validation_42444966_40403.785_00004484
40
+ ARKitScenes_test Validation_42444968_40282.385_00002269
41
+ ARKitScenes_test Validation_42444968_40298.395_00003230
42
+ ARKitScenes_test Validation_42444968_40314.388_00004190
43
+ ARKitScenes_test Validation_42444968_40317.387_00004370
44
+ ARKitScenes_test Validation_42444976_40167.383_00001597
45
+ ARKitScenes_test Validation_42444976_40181.394_00002438
46
+ ARKitScenes_test Validation_42444976_40230.390_00005379
47
+ ARKitScenes_test Validation_42445021_48888.256_00000516
48
+ ARKitScenes_test Validation_42445026_48955.063_00001189
49
+ ARKitScenes_test Validation_42445031_49451.392_00000403
50
+ ARKitScenes_test Validation_42445031_49460.389_00000943
51
+ ARKitScenes_test Validation_42445031_49507.387_00003764
52
+ ARKitScenes_test Validation_42445031_49509.403_00003885
53
+ ARKitScenes_test Validation_42445031_49515.400_00004245
54
+ ARKitScenes_test Validation_42445429_50317.089_00000037
55
+ ARKitScenes_test Validation_42445441_50215.779_00000096
56
+ ARKitScenes_test Validation_42445441_50225.792_00000697
57
+ ARKitScenes_test Validation_42445448_50409.603_00000056
58
+ ARKitScenes_test Validation_42445448_50484.590_00004557
59
+ ARKitScenes_test Validation_42445448_50505.581_00005817
60
+ ARKitScenes_test Validation_42445991_75294.425_00002323
61
+ ARKitScenes_test Validation_42445991_75309.436_00003224
62
+ ARKitScenes_test Validation_42445991_75325.429_00004184
63
+ ARKitScenes_test Validation_42446038_347874.972_00000943
64
+ ARKitScenes_test Validation_42446049_347782.076_00000517
65
+ ARKitScenes_test Validation_42446100_77922.291_00000937
66
+ ARKitScenes_test Validation_42446100_77925.290_00001117
67
+ ARKitScenes_test Validation_42446100_77951.280_00002677
68
+ ARKitScenes_test Validation_42446100_77954.295_00002858
69
+ ARKitScenes_test Validation_42446100_77956.294_00002978
70
+ ARKitScenes_test Validation_42446100_77967.290_00003638
71
+ ARKitScenes_test Validation_42446100_77971.288_00003878
72
+ ARKitScenes_test Validation_42446100_77978.285_00004298
73
+ ARKitScenes_test Validation_42446100_78000.293_00005619
74
+ ARKitScenes_test Validation_42446100_78004.292_00005859
75
+ ARKitScenes_test Validation_42446100_78012.289_00006339
76
+ ARKitScenes_test Validation_42446100_78018.286_00006699
77
+ ARKitScenes_test Validation_42446100_78021.285_00006879
78
+ ARKitScenes_test Validation_42446100_78025.284_00007119
79
+ ARKitScenes_test Validation_42446103_78053.189_00000043
80
+ ARKitScenes_test Validation_42446103_78059.187_00000403
81
+ ARKitScenes_test Validation_42446103_78098.188_00002744
82
+ ARKitScenes_test Validation_42446103_78113.182_00003644
83
+ ARKitScenes_test Validation_42446103_78120.196_00004065
84
+ ARKitScenes_test Validation_42446103_78124.194_00004305
85
+ ARKitScenes_test Validation_42446103_78125.194_00004365
86
+ ARKitScenes_test Validation_42446103_78131.192_00004725
87
+ ARKitScenes_test Validation_42446103_78165.195_00006766
88
+ ARKitScenes_test Validation_42446103_78170.193_00007066
89
+ ARKitScenes_test Validation_42446103_78176.191_00007426
90
+ ARKitScenes_test Validation_42446114_78209.994_00001298
91
+ ARKitScenes_test Validation_42446114_78218.990_00001838
92
+ ARKitScenes_test Validation_42446114_78220.989_00001958
93
+ ARKitScenes_test Validation_42446114_78221.989_00002018
94
+ ARKitScenes_test Validation_42446114_78238.982_00003038
95
+ ARKitScenes_test Validation_42446114_78241.981_00003218
96
+ ARKitScenes_test Validation_42446114_78258.991_00004239
97
+ ARKitScenes_test Validation_42446114_78289.995_00006100
98
+ ARKitScenes_test Validation_42446114_78314.986_00007600
99
+ ARKitScenes_test Validation_42446114_78322.982_00008080
100
+ ARKitScenes_test Validation_42446114_78327.997_00008381
101
+ ARKitScenes_test Validation_42446156_79145.119_00000577
102
+ ARKitScenes_test Validation_42446163_79567.523_00000097
103
+ ARKitScenes_test Validation_42446163_79575.520_00000577
104
+ ARKitScenes_test Validation_42446165_79787.719_00004358
105
+ ARKitScenes_test Validation_42446167_79666.734_00001424
106
+ ARKitScenes_test Validation_42446167_79674.731_00001904
107
+ ARKitScenes_test Validation_42446167_79675.730_00001964
108
+ ARKitScenes_test Validation_42446167_79693.723_00003044
109
+ ARKitScenes_test Validation_42446167_79708.734_00003945
110
+ ARKitScenes_test Validation_42446517_197446.270_00001717
111
+ ARKitScenes_test Validation_42446517_197455.283_00002258
112
+ ARKitScenes_test Validation_42446519_197493.884_00000578
113
+ ARKitScenes_test Validation_42446519_197515.875_00001898
114
+ ARKitScenes_test Validation_42446519_197554.875_00004239
115
+ ARKitScenes_test Validation_42446519_197555.875_00004299
116
+ ARKitScenes_test Validation_42446519_197593.876_00006580
117
+ ARKitScenes_test Validation_42446522_197357.873_00002647
118
+ ARKitScenes_test Validation_42446522_197389.877_00004568
119
+ ARKitScenes_test Validation_42446522_197392.875_00004748
120
+ ARKitScenes_test Validation_42446522_197394.874_00004868
121
+ ARKitScenes_test Validation_42446527_198630.850_00000301
122
+ ARKitScenes_test Validation_42446529_198723.845_00000133
123
+ ARKitScenes_test Validation_42446529_198742.837_00001273
124
+ ARKitScenes_test Validation_42446529_198770.842_00002954
125
+ ARKitScenes_test Validation_42446532_198850.942_00002600
126
+ ARKitScenes_test Validation_42446532_198871.950_00003861
127
+ ARKitScenes_test Validation_42446533_200402.294_00000535
128
+ ARKitScenes_test Validation_42446533_200404.294_00000655
129
+ ARKitScenes_test Validation_42446535_200530.492_00003188
130
+ ARKitScenes_test Validation_42446540_201462.348_00000691
131
+ ARKitScenes_test Validation_42446540_201472.344_00001291
132
+ ARKitScenes_test Validation_42446540_201478.342_00001651
133
+ ARKitScenes_test Validation_42446540_201482.340_00001891
134
+ ARKitScenes_test Validation_42446540_201489.337_00002311
135
+ ARKitScenes_test Validation_42446540_201499.350_00002912
136
+ ARKitScenes_test Validation_42446540_201502.348_00003092
137
+ ARKitScenes_test Validation_42446540_201517.342_00003992
138
+ ARKitScenes_test Validation_42446540_201524.339_00004412
139
+ ARKitScenes_test Validation_42446540_201526.338_00004532
140
+ ARKitScenes_test Validation_42446540_201534.335_00005012
141
+ ARKitScenes_test Validation_42446540_201536.351_00005133
142
+ ARKitScenes_test Validation_42446541_201563.440_00000943
143
+ ARKitScenes_test Validation_42446541_201567.438_00001183
144
+ ARKitScenes_test Validation_42446541_201628.446_00004845
145
+ ARKitScenes_test Validation_42446541_201636.443_00005325
146
+ ARKitScenes_test Validation_42446541_201638.442_00005445
147
+ ARKitScenes_test Validation_42446541_201657.434_00006585
148
+ ARKitScenes_test Validation_42446541_201661.433_00006825
149
+ ARKitScenes_test Validation_42897501_270032.593_00001466
150
+ ARKitScenes_test Validation_42897501_270048.586_00002426
151
+ ARKitScenes_test Validation_42897501_270050.586_00002546
152
+ ARKitScenes_test Validation_42897501_270053.584_00002726
153
+ ARKitScenes_test Validation_42897501_270069.578_00003686
154
+ ARKitScenes_test Validation_42897504_269846.486_00000259
155
+ ARKitScenes_test Validation_42897504_269883.487_00002480
156
+ ARKitScenes_test Validation_42897504_269889.485_00002840
157
+ ARKitScenes_test Validation_42897504_269902.479_00003620
158
+ ARKitScenes_test Validation_42897508_269962.088_00002384
159
+ ARKitScenes_test Validation_42897508_269975.083_00003164
160
+ ARKitScenes_test Validation_42897521_273988.900_00002162
161
+ ARKitScenes_test Validation_42897526_273865.400_00001255
162
+ ARKitScenes_test Validation_42897526_273873.397_00001735
163
+ ARKitScenes_test Validation_42897528_273919.095_00001141
164
+ ARKitScenes_test Validation_42897528_273940.103_00002402
165
+ ARKitScenes_test Validation_42897541_275396.527_00001285
166
+ ARKitScenes_test Validation_42897541_275404.540_00001766
167
+ ARKitScenes_test Validation_42897541_275436.527_00003686
168
+ ARKitScenes_test Validation_42897542_275468.131_00000811
169
+ ARKitScenes_test Validation_42897542_275483.141_00001712
170
+ ARKitScenes_test Validation_42897549_276508.368_00001387
171
+ ARKitScenes_test Validation_42897549_276512.366_00001627
172
+ ARKitScenes_test Validation_42897552_277361.813_00000217
173
+ ARKitScenes_test Validation_42897554_277333.008_00001135
174
+ ARKitScenes_test Validation_42897559_277434.600_00000241
175
+ ARKitScenes_test Validation_42897561_278930.716_00001159
176
+ ARKitScenes_test Validation_42897561_278945.710_00002059
177
+ ARKitScenes_test Validation_42897561_278946.710_00002119
178
+ ARKitScenes_test Validation_42897561_278950.708_00002359
179
+ ARKitScenes_test Validation_42897566_279057.014_00001076
180
+ ARKitScenes_test Validation_42897566_279079.005_00002396
181
+ ARKitScenes_test Validation_42897599_476331.200_00001789
182
+ ARKitScenes_test Validation_42897599_476342.212_00002450
183
+ ARKitScenes_test Validation_42897599_476351.209_00002990
184
+ ARKitScenes_test Validation_42897599_476408.202_00006411
185
+ ARKitScenes_test Validation_42897599_476442.205_00008452
186
+ ARKitScenes_test Validation_42897599_476474.208_00010373
187
+ ARKitScenes_test Validation_42897599_476495.200_00011633
188
+ ARKitScenes_test Validation_42897599_476498.199_00011813
189
+ ARKitScenes_test Validation_42897599_476518.207_00013014
190
+ ARKitScenes_test Validation_42897599_476522.205_00013254
191
+ ARKitScenes_test Validation_42897599_476559.207_00015475
192
+ ARKitScenes_test Validation_42897647_478140.187_00000643
193
+ ARKitScenes_test Validation_42897647_478179.188_00002984
194
+ ARKitScenes_test Validation_42897647_478188.184_00003524
195
+ ARKitScenes_test Validation_42897647_478200.179_00004244
196
+ ARKitScenes_test Validation_42897647_478203.178_00004424
197
+ ARKitScenes_test Validation_42897647_478207.176_00004664
198
+ ARKitScenes_test Validation_42897647_478215.190_00005145
199
+ ARKitScenes_test Validation_42897651_478324.278_00004796
200
+ ARKitScenes_test Validation_42897651_478325.278_00004856
201
+ ARKitScenes_test Validation_42897651_478334.274_00005396
202
+ ARKitScenes_test Validation_42897651_478344.286_00005997
203
+ ARKitScenes_test Validation_42897651_478354.282_00006597
204
+ ARKitScenes_test Validation_42897651_478357.281_00006777
205
+ ARKitScenes_test Validation_42897651_478370.276_00007557
206
+ ARKitScenes_test Validation_42897651_478433.283_00011339
207
+ ARKitScenes_test Validation_42897651_478439.281_00011699
208
+ ARKitScenes_test Validation_42897651_478440.280_00011759
209
+ ARKitScenes_test Validation_42897651_478458.273_00012839
210
+ ARKitScenes_test Validation_42897651_478463.271_00013139
211
+ ARKitScenes_test Validation_42897651_478473.283_00013740
212
+ ARKitScenes_test Validation_42897651_478477.282_00013980
213
+ ARKitScenes_test Validation_42897651_478481.280_00014220
214
+ ARKitScenes_test Validation_42897651_478482.280_00014280
215
+ ARKitScenes_test Validation_42897651_478498.273_00015240
216
+ ARKitScenes_test Validation_42897667_478043.077_00002797
217
+ ARKitScenes_test Validation_42897667_478052.090_00003338
218
+ ARKitScenes_test Validation_42897667_478080.078_00005018
219
+ ARKitScenes_test Validation_42897667_478084.077_00005258
220
+ ARKitScenes_test Validation_42897667_478120.079_00007419
221
+ ARKitScenes_test Validation_42897688_479152.078_00000516
222
+ ARKitScenes_test Validation_42897688_479155.076_00000696
223
+ ARKitScenes_test Validation_42897688_479181.082_00002257
224
+ ARKitScenes_test Validation_42897688_479189.079_00002737
225
+ ARKitScenes_test Validation_42897688_479227.080_00005018
226
+ ARKitScenes_test Validation_42897688_479273.078_00007779
227
+ ARKitScenes_test Validation_42897688_479276.077_00007959
228
+ ARKitScenes_test Validation_42897688_479302.082_00009520
229
+ ARKitScenes_test Validation_42897688_479303.082_00009580
230
+ ARKitScenes_test Validation_42898811_167554.386_00005018
231
+ ARKitScenes_test Validation_42898816_167393.902_00002444
232
+ ARKitScenes_test Validation_42898818_168178.187_00001598
233
+ ARKitScenes_test Validation_42898822_168207.574_00000157
234
+ ARKitScenes_test Validation_42898822_168249.574_00002678
235
+ ARKitScenes_test Validation_42898822_168262.585_00003459
236
+ ARKitScenes_test Validation_42898826_168099.186_00000103
237
+ ARKitScenes_test Validation_42898849_196451.916_00000397
238
+ ARKitScenes_test Validation_42898849_196462.928_00001058
239
+ ARKitScenes_test Validation_42898849_196469.926_00001478
240
+ ARKitScenes_test Validation_42898849_196475.923_00001838
241
+ ARKitScenes_test Validation_42898849_196483.920_00002318
242
+ ARKitScenes_test Validation_42898849_196484.920_00002378
243
+ ARKitScenes_test Validation_42898849_196497.915_00003158
244
+ ARKitScenes_test Validation_42898849_196517.923_00004359
245
+ ARKitScenes_test Validation_42898849_196543.913_00005919
246
+ ARKitScenes_test Validation_42898849_196544.913_00005979
247
+ ARKitScenes_test Validation_42898849_196546.928_00006100
248
+ ARKitScenes_test Validation_42898849_196563.922_00007120
249
+ ARKitScenes_test Validation_42898849_196582.914_00008260
250
+ ARKitScenes_test Validation_42898854_196214.727_00001928
251
+ ARKitScenes_test Validation_42898854_196218.725_00002168
252
+ ARKitScenes_test Validation_42898854_196260.725_00004689
253
+ ARKitScenes_test Validation_42898854_196276.719_00005649
254
+ ARKitScenes_test Validation_42898854_196281.717_00005949
255
+ ARKitScenes_test Validation_42898854_196289.714_00006429
256
+ ARKitScenes_test Validation_42898854_196292.712_00006609
257
+ ARKitScenes_test Validation_42898854_196297.727_00006910
258
+ ARKitScenes_test Validation_42898862_196317.719_00000391
259
+ ARKitScenes_test Validation_42898862_196324.716_00000811
260
+ ARKitScenes_test Validation_42898862_196328.715_00001051
261
+ ARKitScenes_test Validation_42898862_196330.714_00001171
262
+ ARKitScenes_test Validation_42898862_196343.726_00001952
263
+ ARKitScenes_test Validation_42898862_196348.724_00002252
264
+ ARKitScenes_test Validation_42898862_196352.722_00002492
265
+ ARKitScenes_test Validation_42898862_196361.718_00003032
266
+ ARKitScenes_test Validation_42898862_196364.717_00003212
267
+ ARKitScenes_test Validation_42898862_196371.714_00003632
268
+ ARKitScenes_test Validation_42898862_196382.727_00004293
269
+ ARKitScenes_test Validation_42898862_196385.726_00004473
270
+ ARKitScenes_test Validation_42898862_196387.725_00004593
271
+ ARKitScenes_test Validation_42898862_196396.721_00005133
272
+ ARKitScenes_test Validation_42898862_196418.713_00006453
273
+ ARKitScenes_test Validation_42898862_196423.727_00006754
274
+ ARKitScenes_test Validation_42899461_195801.429_00000997
275
+ ARKitScenes_test Validation_42899461_195803.428_00001117
276
+ ARKitScenes_test Validation_42899461_195807.426_00001357
277
+ ARKitScenes_test Validation_42899461_195865.419_00004838
278
+ ARKitScenes_test Validation_42899461_195869.434_00005079
279
+ ARKitScenes_test Validation_42899461_195881.429_00005799
280
+ ARKitScenes_test Validation_42899471_195673.532_00000938
281
+ ARKitScenes_test Validation_42899471_195705.535_00002859
282
+ ARKitScenes_test Validation_42899471_195706.535_00002919
283
+ ARKitScenes_test Validation_42899471_195710.533_00003159
284
+ ARKitScenes_test Validation_42899611_228365.713_00000997
285
+ ARKitScenes_test Validation_42899611_228379.724_00001838
286
+ ARKitScenes_test Validation_42899611_228387.721_00002318
287
+ ARKitScenes_test Validation_42899611_228400.716_00003098
288
+ ARKitScenes_test Validation_42899611_228401.716_00003158
289
+ ARKitScenes_test Validation_42899611_228411.712_00003758
290
+ ARKitScenes_test Validation_42899612_228429.521_00000229
291
+ ARKitScenes_test Validation_42899612_228433.520_00000469
292
+ ARKitScenes_test Validation_42899612_228469.522_00002630
293
+ ARKitScenes_test Validation_42899612_228476.519_00003050
294
+ ARKitScenes_test Validation_42899612_228479.518_00003230
295
+ ARKitScenes_test Validation_42899612_228481.517_00003350
296
+ ARKitScenes_test Validation_42899612_228503.525_00004671
297
+ ARKitScenes_test Validation_42899612_228515.521_00005391
298
+ ARKitScenes_test Validation_42899613_228274.016_00000102
299
+ ARKitScenes_test Validation_42899613_228311.018_00002323
300
+ ARKitScenes_test Validation_42899613_228315.017_00002563
301
+ ARKitScenes_test Validation_42899619_228890.872_00000157
302
+ ARKitScenes_test Validation_42899619_228899.869_00000697
303
+ ARKitScenes_test Validation_42899620_228848.672_00000097
304
+ ARKitScenes_test Validation_44358442_42896.098_00000343
305
+ ARKitScenes_test Validation_44358442_42911.092_00001243
306
+ ARKitScenes_test Validation_44358446_43064.895_00003398
307
+ ARKitScenes_test Validation_44358446_43120.888_00006759
308
+ ARKitScenes_test Validation_44358448_42965.802_00000955
309
+ ARKitScenes_test Validation_44358448_42978.797_00001735
310
+ ARKitScenes_test Validation_44358448_42996.790_00002815
311
+ ARKitScenes_test Validation_44358448_42998.789_00002935
312
+ ARKitScenes_test Validation_44358448_43000.805_00003056
313
+ ARKitScenes_test Validation_44358451_43608.890_00000217
314
+ ARKitScenes_test Validation_44358451_43610.889_00000337
315
+ ARKitScenes_test Validation_44358451_43614.888_00000577
316
+ ARKitScenes_test Validation_44358451_43619.886_00000877
317
+ ARKitScenes_test Validation_44358451_43640.894_00002138
318
+ ARKitScenes_test Validation_44358451_43669.881_00003878
319
+ ARKitScenes_test Validation_44358451_43714.896_00006580
320
+ ARKitScenes_test Validation_44358452_43508.598_00000517
321
+ ARKitScenes_test Validation_44358452_43509.598_00000577
322
+ ARKitScenes_test Validation_44358452_43586.583_00005198
323
+ ARKitScenes_test Validation_44358452_43590.598_00005439
324
+ ARKitScenes_test Validation_45260854_575.085_00001129
325
+ ARKitScenes_test Validation_45260854_578.084_00001309
326
+ ARKitScenes_test Validation_45260854_593.078_00002209
327
+ ARKitScenes_test Validation_45260854_604.090_00002870
328
+ ARKitScenes_test Validation_45260856_632.278_00000396
329
+ ARKitScenes_test Validation_45260856_635.277_00000576
330
+ ARKitScenes_test Validation_45260856_656.285_00001837
331
+ ARKitScenes_test Validation_45260856_694.286_00004118
332
+ ARKitScenes_test Validation_45260856_707.281_00004898
333
+ ARKitScenes_test Validation_45260856_713.278_00005258
334
+ ARKitScenes_test Validation_45260857_501.682_00000277
335
+ ARKitScenes_test Validation_45260857_509.679_00000757
336
+ ARKitScenes_test Validation_45260857_515.693_00001118
337
+ ARKitScenes_test Validation_45260857_518.692_00001298
338
+ ARKitScenes_test Validation_45260857_533.686_00002198
339
+ ARKitScenes_test Validation_45260857_542.682_00002738
340
+ ARKitScenes_test Validation_45260898_786.983_00001429
341
+ ARKitScenes_test Validation_45260899_887.793_00002384
342
+ ARKitScenes_test Validation_45260899_896.789_00002924
343
+ ARKitScenes_test Validation_45260900_833.481_00001813
344
+ ARKitScenes_test Validation_45260920_2470.417_00000228
345
+ ARKitScenes_test Validation_45260920_2475.432_00000529
346
+ ARKitScenes_test Validation_45260920_2493.425_00001609
347
+ ARKitScenes_test Validation_45260920_2508.419_00002509
348
+ ARKitScenes_test Validation_45260920_2534.425_00004070
349
+ ARKitScenes_test Validation_45260925_2396.830_00000157
350
+ ARKitScenes_test Validation_45260925_2429.817_00002137
351
+ ARKitScenes_test Validation_45260925_2432.816_00002317
352
+ ARKitScenes_test Validation_45260925_2434.831_00002438
353
+ ARKitScenes_test Validation_45260928_2379.220_00002498
354
+ ARKitScenes_test Validation_45261121_3431.070_00001034
355
+ ARKitScenes_test Validation_45261121_3433.069_00001154
356
+ ARKitScenes_test Validation_45261121_3445.064_00001874
357
+ ARKitScenes_test Validation_45261121_3446.063_00001934
358
+ ARKitScenes_test Validation_45261121_3449.062_00002114
359
+ ARKitScenes_test Validation_45261121_3451.061_00002234
360
+ ARKitScenes_test Validation_45261121_3455.060_00002474
361
+ ARKitScenes_test Validation_45261121_3460.058_00002774
362
+ ARKitScenes_test Validation_45261128_3497.159_00000457
363
+ ARKitScenes_test Validation_45261128_3500.158_00000637
364
+ ARKitScenes_test Validation_45261128_3502.157_00000757
365
+ ARKitScenes_test Validation_45261128_3534.160_00002678
366
+ ARKitScenes_test Validation_45261133_4053.263_00002797
367
+ ARKitScenes_test Validation_45261133_4093.263_00005198
368
+ ARKitScenes_test Validation_45261133_4098.261_00005498
369
+ ARKitScenes_test Validation_45261133_4108.273_00006099
370
+ ARKitScenes_test Validation_45261133_4134.262_00007659
371
+ ARKitScenes_test Validation_45261133_4135.262_00007719
372
+ ARKitScenes_test Validation_45261133_4139.260_00007959
373
+ ARKitScenes_test Validation_45261133_4141.259_00008079
374
+ ARKitScenes_test Validation_45261140_3956.270_00000583
375
+ ARKitScenes_test Validation_45261140_3958.269_00000703
376
+ ARKitScenes_test Validation_45261140_3977.261_00001843
377
+ ARKitScenes_test Validation_45261140_3985.274_00002324
378
+ ARKitScenes_test Validation_45261142_3924.466_00002335
379
+ ARKitScenes_test Validation_45261143_4589.702_00000643
380
+ ARKitScenes_test Validation_45261143_4602.714_00001424
381
+ ARKitScenes_test Validation_45261143_4630.702_00003104
382
+ ARKitScenes_test Validation_45261144_4656.008_00000337
383
+ ARKitScenes_test Validation_45261150_4553.301_00001657
384
+ ARKitScenes_test Validation_45261182_2970.729_00003398
385
+ ARKitScenes_test Validation_45261182_2972.728_00003518
386
+ ARKitScenes_test Validation_45261185_3551.484_00000283
387
+ ARKitScenes_test Validation_45261185_3561.480_00000883
388
+ ARKitScenes_test Validation_45261185_3581.489_00002084
389
+ ARKitScenes_test Validation_45261190_3699.892_00000823
390
+ ARKitScenes_test Validation_45261190_3709.888_00001423
391
+ ARKitScenes_test Validation_45261190_3735.895_00002984
392
+ ARKitScenes_test Validation_45261193_3621.190_00000169
393
+ ARKitScenes_test Validation_45261193_3646.180_00001669
394
+ ARKitScenes_test Validation_45261193_3654.194_00002150
395
+ ARKitScenes_test Validation_45261575_77494.516_00001538
396
+ ARKitScenes_test Validation_45261581_77517.507_00000097
397
+ ARKitScenes_test Validation_45261581_77537.515_00001298
398
+ ARKitScenes_test Validation_45261581_77538.515_00001358
399
+ ARKitScenes_test Validation_45261581_77539.514_00001418
400
+ ARKitScenes_test Validation_45261581_77560.505_00002678
401
+ ARKitScenes_test Validation_45261581_77587.511_00004299
402
+ ARKitScenes_test Validation_45261581_77590.510_00004479
403
+ ARKitScenes_test Validation_45261588_78074.609_00000637
404
+ ARKitScenes_test Validation_45261588_78081.606_00001057
405
+ ARKitScenes_test Validation_45261588_78083.605_00001177
406
+ ARKitScenes_test Validation_45261588_78091.602_00001657
407
+ ARKitScenes_test Validation_45261594_78053.201_00001477
408
+ ARKitScenes_test Validation_45261619_55117.682_00003176
409
+ ARKitScenes_test Validation_45261620_54953.280_00001297
410
+ ARKitScenes_test Validation_45261631_56074.403_00002858
411
+ ARKitScenes_test Validation_45261631_56077.402_00003038
412
+ ARKitScenes_test Validation_45261632_55901.188_00000678
413
+ ARKitScenes_test Validation_45261637_55970.494_00000877
414
+ ARKitScenes_test Validation_45662921_21777.915_00001627
415
+ ARKitScenes_test Validation_45662921_21785.928_00002108
416
+ ARKitScenes_test Validation_45662921_21792.926_00002528
417
+ ARKitScenes_test Validation_45662921_21795.924_00002708
418
+ ARKitScenes_test Validation_45662921_21830.926_00004809
419
+ ARKitScenes_test Validation_45662921_21843.921_00005589
420
+ ARKitScenes_test Validation_45662921_21844.921_00005649
421
+ ARKitScenes_test Validation_45662921_21852.917_00006129
422
+ ARKitScenes_test Validation_45662924_21565.920_00000637
423
+ ARKitScenes_test Validation_45662924_21575.916_00001237
424
+ ARKitScenes_test Validation_45662924_21603.921_00002918
425
+ ARKitScenes_test Validation_45662926_21672.326_00002048
426
+ ARKitScenes_test Validation_45662926_21678.323_00002408
427
+ ARKitScenes_test Validation_45662926_21679.323_00002468
428
+ ARKitScenes_test Validation_45662926_21701.330_00003789
429
+ ARKitScenes_test Validation_45662926_21718.323_00004809
430
+ ARKitScenes_test Validation_45662926_21719.323_00004869
431
+ ARKitScenes_test Validation_45662926_21720.322_00004929
432
+ ARKitScenes_test Validation_45662942_22882.675_00000037
433
+ ARKitScenes_test Validation_45662942_22899.668_00001057
434
+ ARKitScenes_test Validation_45662942_22915.661_00002017
435
+ ARKitScenes_test Validation_45662942_22930.672_00002918
436
+ ARKitScenes_test Validation_45662943_23021.267_00000157
437
+ ARKitScenes_test Validation_45662943_23140.268_00007300
438
+ ARKitScenes_test Validation_45662943_23172.254_00009220
439
+ ARKitScenes_test Validation_45662943_23198.260_00010781
440
+ ARKitScenes_test Validation_45662943_23200.259_00010901
441
+ ARKitScenes_test Validation_45662943_23203.258_00011081
442
+ ARKitScenes_test Validation_45662944_22955.761_00000817
443
+ ARKitScenes_test Validation_45662970_26110.534_00000247
444
+ ARKitScenes_test Validation_45662970_26162.530_00003368
445
+ ARKitScenes_test Validation_45662970_26164.529_00003488
446
+ ARKitScenes_test Validation_45662970_26169.527_00003788
447
+ ARKitScenes_test Validation_45662970_26170.527_00003848
448
+ ARKitScenes_test Validation_45662975_26067.134_00000163
449
+ ARKitScenes_test Validation_45662975_26074.131_00000583
450
+ ARKitScenes_test Validation_45662981_27941.098_00000289
451
+ ARKitScenes_test Validation_45662981_28034.095_00005871
452
+ ARKitScenes_test Validation_45662983_27797.888_00000097
453
+ ARKitScenes_test Validation_45662983_27799.887_00000217
454
+ ARKitScenes_test Validation_45662983_27803.886_00000457
455
+ ARKitScenes_test Validation_45662987_27872.192_00000349
456
+ ARKitScenes_test Validation_45663113_55333.668_00001778
457
+ ARKitScenes_test Validation_45663114_55382.864_00000397
458
+ ARKitScenes_test Validation_45663114_55385.863_00000577
459
+ ARKitScenes_test Validation_45663114_55413.868_00002258
460
+ ARKitScenes_test Validation_45663114_55415.867_00002378
461
+ ARKitScenes_test Validation_45663115_55471.860_00002558
462
+ ARKitScenes_test Validation_45663149_63308.707_00000757
463
+ ARKitScenes_test Validation_45663149_63312.706_00000997
464
+ ARKitScenes_test Validation_45663150_63508.953_00000108
465
+ ARKitScenes_test Validation_45663150_63515.950_00000528
466
+ ARKitScenes_test Validation_45663150_63519.949_00000768
467
+ ARKitScenes_test Validation_45663150_63548.954_00002509
468
+ ARKitScenes_test Validation_45663150_63575.960_00004130
469
+ ARKitScenes_test Validation_45663150_63579.958_00004370
470
+ ARKitScenes_test Validation_45663150_63582.957_00004550
471
+ ARKitScenes_test Validation_45663154_63400.863_00000133
472
+ ARKitScenes_test Validation_45663154_63407.860_00000553
473
+ ARKitScenes_test Validation_45663154_63408.859_00000613
474
+ ARKitScenes_test Validation_45663154_63435.849_00002233
475
+ ARKitScenes_test Validation_45663154_63436.848_00002293
476
+ ARKitScenes_test Validation_45663154_63451.859_00003194
477
+ ARKitScenes_test Validation_45663164_65934.817_00000234
478
+ ARKitScenes_test Validation_45663164_65937.816_00000414
479
+ ARKitScenes_test Validation_45663164_65954.825_00001435
480
+ ARKitScenes_test Validation_45663164_65965.821_00002095
481
+ ARKitScenes_test Validation_45663164_65971.819_00002455
482
+ ARKitScenes_test Validation_45663164_65981.815_00003055
483
+ ARKitScenes_test Validation_45663164_65990.828_00003596
484
+ ARKitScenes_test Validation_45663164_66002.823_00004316
485
+ ARKitScenes_test Validation_45663165_65825.027_00001238
486
+ ARKitScenes_test Validation_45663165_65826.026_00001298
487
+ ARKitScenes_test Validation_45663175_65888.618_00001033
488
+ ARKitScenes_test Validation_45663175_65895.616_00001453
489
+ ARKitScenes_test Validation_45663175_65900.614_00001753
490
+ ARKitScenes_test Validation_47115452_13059.567_00000217
491
+ ARKitScenes_test Validation_47115452_13068.580_00000758
492
+ ARKitScenes_test Validation_47115452_13087.572_00001898
493
+ ARKitScenes_test Validation_47115452_13111.579_00003339
494
+ ARKitScenes_test Validation_47115452_13118.576_00003759
495
+ ARKitScenes_test Validation_47115452_13119.575_00003819
496
+ ARKitScenes_test Validation_47115460_13170.771_00002029
497
+ ARKitScenes_test Validation_47115463_13276.477_00003458
498
+ ARKitScenes_test Validation_47115463_13302.466_00005018
499
+ ARKitScenes_test Validation_47115469_14015.292_00000277
500
+ ARKitScenes_test Validation_47115469_14028.286_00001057
501
+ ARKitScenes_test Validation_47115469_14030.285_00001177
502
+ ARKitScenes_test Validation_47115473_14110.785_00002330
503
+ ARKitScenes_test Validation_47115473_14130.777_00003530
504
+ ARKitScenes_test Validation_47115473_14143.788_00004311
505
+ ARKitScenes_test Validation_47115473_14173.776_00006111
506
+ ARKitScenes_test Validation_47115474_13964.679_00000282
507
+ ARKitScenes_test Validation_47115474_13981.689_00001303
508
+ ARKitScenes_test Validation_47204552_3362.090_00000577
509
+ ARKitScenes_test Validation_47204552_3393.094_00002438
510
+ ARKitScenes_test Validation_47204552_3399.091_00002798
511
+ ARKitScenes_test Validation_47204552_3401.091_00002918
512
+ ARKitScenes_test Validation_47204552_3402.090_00002978
513
+ ARKitScenes_test Validation_47204552_3415.101_00003759
514
+ ARKitScenes_test Validation_47204552_3423.098_00004239
515
+ ARKitScenes_test Validation_47204552_3427.097_00004479
516
+ ARKitScenes_test Validation_47204552_3438.092_00005139
517
+ ARKitScenes_test Validation_47204552_3450.087_00005859
518
+ ARKitScenes_test Validation_47204552_3478.092_00007540
519
+ ARKitScenes_test Validation_47204554_3222.198_00000697
520
+ ARKitScenes_test Validation_47204554_3262.198_00003098
521
+ ARKitScenes_test Validation_47204554_3276.192_00003938
522
+ ARKitScenes_test Validation_47204556_3287.388_00000096
523
+ ARKitScenes_test Validation_47204556_3289.387_00000216
524
+ ARKitScenes_test Validation_47204556_3296.401_00000637
525
+ ARKitScenes_test Validation_47204556_3318.392_00001957
526
+ ARKitScenes_test Validation_47204556_3339.400_00003218
527
+ ARKitScenes_test Validation_47204559_4004.548_00000997
528
+ ARKitScenes_test Validation_47204559_4005.547_00001057
529
+ ARKitScenes_test Validation_47204563_4035.451_00000157
530
+ ARKitScenes_test Validation_47204566_4079.050_00000217
531
+ ARKitScenes_test Validation_47204566_4108.038_00001957
532
+ ARKitScenes_test Validation_47204566_4112.036_00002197
533
+ ARKitScenes_test Validation_47204566_4122.049_00002798
534
+ ARKitScenes_test Validation_47204566_4127.047_00003098
535
+ ARKitScenes_test Validation_47204573_4957.623_00000038
536
+ ARKitScenes_test Validation_47204573_4997.607_00002438
537
+ ARKitScenes_test Validation_47204573_5033.609_00004599
538
+ ARKitScenes_test Validation_47204573_5037.607_00004839
539
+ ARKitScenes_test Validation_47204575_4843.221_00000529
540
+ ARKitScenes_test Validation_47204575_4872.209_00002269
541
+ ARKitScenes_test Validation_47204575_4874.208_00002389
542
+ ARKitScenes_test Validation_47204578_4911.209_00001063
543
+ ARKitScenes_test Validation_47204578_4929.219_00002144
544
+ ARKitScenes_test Validation_47204605_2404.299_00000121
545
+ ARKitScenes_test Validation_47204605_2412.295_00000601
546
+ ARKitScenes_test Validation_47204605_2419.293_00001021
547
+ ARKitScenes_test Validation_47204605_2420.292_00001081
548
+ ARKitScenes_test Validation_47204605_2431.305_00001742
549
+ ARKitScenes_test Validation_47204605_2436.303_00002042
550
+ ARKitScenes_test Validation_47204605_2440.301_00002282
551
+ ARKitScenes_test Validation_47204607_2528.799_00000517
552
+ ARKitScenes_test Validation_47204607_2560.803_00002438
553
+ ARKitScenes_test Validation_47204607_2561.803_00002498
554
+ ARKitScenes_test Validation_47204607_2569.800_00002978
555
+ ARKitScenes_test Validation_47204607_2573.798_00003218
556
+ ARKitScenes_test Validation_47204607_2574.798_00003278
557
+ ARKitScenes_test Validation_47204607_2577.797_00003458
558
+ ARKitScenes_test Validation_47204607_2583.794_00003818
559
+ ARKitScenes_test Validation_47204607_2591.791_00004298
560
+ ARKitScenes_test Validation_47204607_2626.794_00006399
561
+ ARKitScenes_test Validation_47204607_2634.791_00006879
562
+ ARKitScenes_test Validation_47204607_2664.796_00008680
563
+ ARKitScenes_test Validation_47204609_2453.013_00000038
564
+ ARKitScenes_test Validation_47204609_2485.000_00001958
565
+ ARKitScenes_test Validation_47331068_131078.065_00000817
566
+ ARKitScenes_test Validation_47331069_131220.357_00004598
567
+ ARKitScenes_test Validation_47331071_131007.360_00000817
568
+ ARKitScenes_test Validation_47331265_90013.317_00000134
569
+ ARKitScenes_test Validation_47331265_90032.293_00001273
570
+ ARKitScenes_test Validation_47331265_90067.295_00003374
571
+ ARKitScenes_test Validation_47331266_89823.527_00000141
572
+ ARKitScenes_test Validation_47331266_89825.493_00000259
573
+ ARKitScenes_test Validation_47331266_89873.490_00003140
574
+ ARKitScenes_test Validation_47331319_179040.028_00000234
575
+ ARKitScenes_test Validation_47331319_179072.032_00002155
576
+ ARKitScenes_test Validation_47331322_179115.348_00000110
577
+ ARKitScenes_test Validation_47331322_179135.339_00001310
578
+ ARKitScenes_test Validation_47331324_179185.452_00000122
579
+ ARKitScenes_test Validation_47331324_179188.434_00000301
580
+ ARKitScenes_test Validation_47331324_179200.430_00001021
581
+ ARKitScenes_test Validation_47331324_179214.424_00001861
582
+ ARKitScenes_test Validation_47331324_179216.440_00001982
583
+ ARKitScenes_test Validation_47331651_22862.230_00000757
584
+ ARKitScenes_test Validation_47331651_22892.235_00002558
585
+ ARKitScenes_test Validation_47331651_22893.235_00002618
586
+ ARKitScenes_test Validation_47331651_22920.224_00004238
587
+ ARKitScenes_test Validation_47331651_22973.235_00007420
588
+ ARKitScenes_test Validation_47331653_23010.337_00001538
589
+ ARKitScenes_test Validation_47331653_23016.334_00001898
590
+ ARKitScenes_test Validation_47331653_23030.329_00002738
591
+ ARKitScenes_test Validation_47331653_23034.327_00002978
592
+ ARKitScenes_test Validation_47331653_23076.326_00005499
593
+ ARKitScenes_test Validation_47331653_23080.325_00005739
594
+ ARKitScenes_test Validation_47331653_23090.337_00006340
595
+ ARKitScenes_test Validation_47331653_23092.337_00006460
596
+ ARKitScenes_test Validation_47331653_23132.320_00008860
597
+ ARKitScenes_test Validation_47331654_22744.828_00000517
598
+ ARKitScenes_test Validation_47331654_22746.827_00000637
599
+ ARKitScenes_test Validation_47331654_22751.825_00000937
600
+ ARKitScenes_test Validation_47331654_22772.834_00002198
601
+ ARKitScenes_test Validation_47331654_22781.830_00002738
602
+ ARKitScenes_test Validation_47331654_22787.827_00003098
603
+ ARKitScenes_test Validation_47331654_22797.823_00003698
604
+ ARKitScenes_test Validation_47331654_22833.825_00005859
605
+ ARKitScenes_test Validation_47331661_23629.582_00002630
606
+ ARKitScenes_test Validation_47331661_23690.590_00006292
607
+ ARKitScenes_test Validation_47331662_23429.080_00000516
608
+ ARKitScenes_test Validation_47331662_23466.082_00002737
609
+ ARKitScenes_test Validation_47331662_23470.080_00002977
610
+ ARKitScenes_test Validation_47331662_23472.079_00003097
611
+ ARKitScenes_test Validation_47331988_170658.548_00003824
612
+ ARKitScenes_test Validation_47331988_170671.559_00004605
613
+ ARKitScenes_test Validation_47331988_170673.559_00004725
614
+ ARKitScenes_test Validation_47331988_170674.558_00004785
615
+ ARKitScenes_test Validation_47331988_170681.555_00005205
616
+ ARKitScenes_test Validation_47331989_170440.853_00001597
617
+ ARKitScenes_test Validation_47331990_170486.951_00000408
618
+ ARKitScenes_test Validation_47331990_170529.950_00002989
619
+ ARKitScenes_test Validation_47331990_170567.951_00005270
620
+ ARKitScenes_test Validation_47331990_170578.963_00005931
621
+ ARKitScenes_test Validation_47332004_171938.432_00000997
622
+ ARKitScenes_test Validation_47332005_172095.735_00001963
623
+ ARKitScenes_test Validation_47332005_172102.732_00002383
624
+ ARKitScenes_test Validation_47332005_172149.729_00005204
625
+ ARKitScenes_test Validation_47332885_7500.859_00001003
626
+ ARKitScenes_test Validation_47332885_7517.869_00002024
627
+ ARKitScenes_test Validation_47332886_7549.256_00000474
628
+ ARKitScenes_test Validation_47332886_7550.256_00000534
629
+ ARKitScenes_test Validation_47332886_7562.267_00001255
630
+ ARKitScenes_test Validation_47332886_7567.265_00001555
631
+ ARKitScenes_test Validation_47332886_7582.259_00002455
632
+ ARKitScenes_test Validation_47332890_7653.364_00003398
633
+ ARKitScenes_test Validation_47332890_7662.360_00003938
634
+ ARKitScenes_test Validation_47332890_7663.360_00003998
635
+ ARKitScenes_test Validation_47332890_7667.358_00004238
636
+ ARKitScenes_test Validation_47332890_7668.358_00004298
637
+ ARKitScenes_test Validation_47332893_9315.687_00007719
638
+ ARKitScenes_test Validation_47332895_9057.792_00000715
639
+ ARKitScenes_test Validation_47332895_9066.788_00001255
640
+ ARKitScenes_test Validation_47332899_9174.194_00003308
641
+ ARKitScenes_test Validation_47332899_9175.194_00003368
642
+ ARKitScenes_test Validation_47332899_9177.193_00003488
643
+ ARKitScenes_test Validation_47332899_9179.192_00003608
644
+ ARKitScenes_test Validation_47332901_11032.594_00000985
645
+ ARKitScenes_test Validation_47332901_11062.598_00002786
646
+ ARKitScenes_test Validation_47332904_11093.386_00000943
647
+ ARKitScenes_test Validation_47332904_11096.385_00001123
648
+ ARKitScenes_test Validation_47332904_11149.396_00004305
649
+ ARKitScenes_test Validation_47332905_10961.889_00000709
650
+ ARKitScenes_test Validation_47332905_10964.888_00000889
651
+ ARKitScenes_test Validation_47332910_12198.325_00000708
652
+ ARKitScenes_test Validation_47332911_12283.823_00001483
653
+ ARKitScenes_test Validation_47332911_12347.831_00005325
654
+ ARKitScenes_test Validation_47332911_12351.829_00005565
655
+ ARKitScenes_test Validation_47332915_13476.969_00000643
656
+ ARKitScenes_test Validation_47332915_13505.974_00002384
657
+ ARKitScenes_test Validation_47332915_13538.977_00004365
658
+ ARKitScenes_test Validation_47332915_13547.973_00004905
659
+ ARKitScenes_test Validation_47332915_13552.971_00005205
660
+ ARKitScenes_test Validation_47332915_13560.968_00005685
661
+ ARKitScenes_test Validation_47332918_13321.866_00000349
662
+ ARKitScenes_test Validation_47332918_13363.865_00002870
663
+ ARKitScenes_test Validation_47333440_53806.225_00000757
664
+ ARKitScenes_test Validation_47333443_53755.129_00000397
665
+ ARKitScenes_test Validation_47333443_53760.127_00000697
666
+ ARKitScenes_test Validation_47333452_54966.217_00000883
667
+ ARKitScenes_test Validation_47333452_54993.223_00002504
668
+ ARKitScenes_test Validation_47333452_54995.222_00002624
669
+ ARKitScenes_test Validation_47333456_55028.108_00001237
670
+ ARKitScenes_test Validation_47333456_55030.107_00001357
671
+ ARKitScenes_test Validation_47333456_55035.122_00001658
672
+ ARKitScenes_test Validation_47333456_55058.112_00003038
673
+ ARKitScenes_test Validation_47333456_55067.109_00003578
674
+ ARKitScenes_test Validation_47333898_81325.974_00000397
675
+ ARKitScenes_test Validation_47333898_81346.966_00001657
676
+ ARKitScenes_test Validation_47333898_81372.972_00003218
677
+ ARKitScenes_test Validation_47333898_81377.970_00003518
678
+ ARKitScenes_test Validation_47333898_81378.969_00003578
679
+ ARKitScenes_test Validation_47333899_81255.070_00000877
680
+ ARKitScenes_test Validation_47333899_81271.063_00001837
681
+ ARKitScenes_test Validation_47333899_81306.066_00003938
682
+ ARKitScenes_test Validation_47333904_81416.371_00000877
683
+ ARKitScenes_test Validation_47333904_81425.367_00001417
684
+ ARKitScenes_test Validation_47333904_81432.364_00001837
685
+ ARKitScenes_test Validation_47333916_82430.565_00003759
686
+ ARKitScenes_test Validation_47333916_82435.563_00004059
687
+ ARKitScenes_test Validation_47333918_82512.465_00003404
688
+ ARKitScenes_test Validation_47333918_82533.457_00004664
689
+ ARKitScenes_test Validation_47333920_82642.262_00005856
690
+ ARKitScenes_test Validation_47333920_82666.253_00007296
691
+ ARKitScenes_test Validation_47333923_83604.501_00000396
692
+ ARKitScenes_test Validation_47333925_83573.613_00001964
693
+ ARKitScenes_test Validation_47333927_84005.212_00000944
694
+ ARKitScenes_test Validation_47333927_84013.209_00001424
695
+ ARKitScenes_test Validation_47333927_84020.206_00001844
696
+ ARKitScenes_test Validation_47333927_84025.204_00002144
697
+ ARKitScenes_test Validation_47333927_84028.203_00002324
698
+ ARKitScenes_test Validation_47333931_83940.705_00001658
699
+ ARKitScenes_test Validation_47333931_83943.704_00001838
700
+ ARKitScenes_test Validation_47333932_84109.203_00002798
701
+ ARKitScenes_test Validation_47333932_84184.206_00007300
702
+ ARKitScenes_test Validation_47333934_85556.772_00000469
703
+ ARKitScenes_test Validation_47333934_85572.765_00001429
704
+ ARKitScenes_test Validation_47333934_85623.778_00004491
705
+ ARKitScenes_test Validation_47333934_85636.773_00005271
706
+ ARKitScenes_test Validation_47333934_85638.772_00005391
707
+ ARKitScenes_test Validation_47333934_85646.769_00005871
708
+ ARKitScenes_test Validation_47333934_85659.763_00006651
709
+ ARKitScenes_test Validation_47333937_85498.362_00000540
710
+ ARKitScenes_test Validation_47333937_85501.378_00000721
711
+ ARKitScenes_test Validation_47333937_85528.367_00002341
712
+ ARKitScenes_test Validation_47333937_85541.378_00003122
713
+ ARKitScenes_test Validation_47333937_85543.377_00003242
714
+ ARKitScenes_test Validation_47333940_85418.978_00000517
715
+ ARKitScenes_test Validation_47333940_85420.977_00000637
716
+ ARKitScenes_test Validation_47333940_85422.976_00000757
717
+ ARKitScenes_test Validation_47333940_85438.969_00001717
718
+ ARKitScenes_test Validation_47333940_85480.969_00004238
719
+ ARKitScenes_test Validation_47334105_92943.745_00003650
720
+ ARKitScenes_test Validation_47334105_92981.746_00005931
721
+ ARKitScenes_test Validation_47334105_93005.753_00007372
722
+ ARKitScenes_test Validation_47334239_111152.021_00002918
723
+ ARKitScenes_test Validation_47334240_111223.425_00003818
724
+ ARKitScenes_test Validation_47429912_136038.580_00000937
725
+ ARKitScenes_test Validation_47429912_136057.589_00002078
726
+ ARKitScenes_test Validation_47429922_136625.256_00000577
727
+ ARKitScenes_test Validation_47429971_16341.763_00003758
728
+ ARKitScenes_test Validation_47429995_16689.687_00000229
729
+ ARKitScenes_test Validation_47430024_18796.033_00001657
730
+ ARKitScenes_test Validation_47430026_18833.034_00001297
731
+ ARKitScenes_test Validation_47430026_18844.046_00001958
732
+ ARKitScenes_test Validation_47430033_19430.380_00001363
733
+ ARKitScenes_test Validation_47430036_19503.284_00000277
734
+ ARKitScenes_test Validation_47430043_20580.283_00000156
735
+ ARKitScenes_test Validation_47430045_20645.389_00001177
736
+ ARKitScenes_test Validation_47430047_21106.832_00000937
737
+ ARKitScenes_test Validation_47430047_21107.832_00000997
738
+ ARKitScenes_test Validation_47895341_96434.329_00002918
739
+ ARKitScenes_test Validation_47895348_96489.523_00001837
740
+ ARKitScenes_test Validation_47895348_96510.515_00003097
741
+ ARKitScenes_test Validation_47895348_96535.521_00004598
742
+ ARKitScenes_test Validation_47895348_96543.518_00005078
743
+ ARKitScenes_test Validation_47895350_96587.817_00002197
744
+ ARKitScenes_test Validation_47895350_96593.814_00002557
745
+ ARKitScenes_test Validation_47895350_96598.812_00002857
746
+ ARKitScenes_test Validation_47895350_96692.824_00008500
747
+ ARKitScenes_test Validation_47895355_98785.640_00001898
748
+ ARKitScenes_test Validation_47895355_98786.640_00001958
749
+ ARKitScenes_test Validation_47895355_98811.630_00003458
750
+ ARKitScenes_test Validation_47895355_98813.629_00003578
751
+ ARKitScenes_test Validation_47895355_98817.627_00003818
752
+ ARKitScenes_test Validation_47895355_98873.638_00007180
753
+ ARKitScenes_test Validation_47895355_98884.633_00007840
754
+ ARKitScenes_test Validation_47895355_98890.631_00008200
755
+ ARKitScenes_test Validation_47895355_98898.628_00008680
756
+ ARKitScenes_test Validation_47895355_98911.639_00009461
757
+ ARKitScenes_test Validation_47895355_98953.639_00011982
758
+ ARKitScenes_test Validation_47895364_100176.943_00000457
759
+ ARKitScenes_test Validation_47895364_100304.941_00008140
760
+ ARKitScenes_test Validation_47895365_100109.337_00000637
761
+ ARKitScenes_test Validation_47895556_457844.157_00003123
762
+ ARKitScenes_test Validation_47895783_387008.265_00000337
763
+ ARKitScenes_test Validation_47895783_387020.260_00001057
764
+ ARKitScenes_test Validation_48018367_389470.584_00001538
765
+ ARKitScenes_test Validation_48018368_389568.277_00002090
766
+ ARKitScenes_test Validation_48018368_389569.276_00002150
767
+ ARKitScenes_test Validation_48018368_389589.268_00003350
768
+ ARKitScenes_test Validation_48018372_389499.672_00000577
769
+ ARKitScenes_test Validation_48018379_390028.141_00001958
770
+ ARKitScenes_test Validation_48018559_3343.849_00000157
771
+ ARKitScenes_test Validation_48018560_3398.743_00000397
772
+ ARKitScenes_test Validation_48018560_3417.752_00001538
773
+ ARKitScenes_test Validation_48018566_3878.430_00001303
774
+ ARKitScenes_test Validation_48018571_3900.055_00000140
775
+ ARKitScenes_test Validation_48018572_3834.032_00002288
776
+ ARKitScenes_test Validation_48018730_1763.469_00005739
777
+ ARKitScenes_test Validation_48018732_1944.579_00002060
778
+ ARKitScenes_test Validation_48018732_1949.577_00002360
779
+ ARKitScenes_test Validation_48018732_1965.571_00003320
780
+ ARKitScenes_test Validation_48018732_1994.575_00005061
781
+ ARKitScenes_test Validation_48458484_111570.316_00000037
782
+ ARKitScenes_test Validation_48458484_111592.307_00001357
783
+ ARKitScenes_test Validation_48458484_111596.305_00001597
784
+ ARKitScenes_test Validation_48458650_7065.603_00000396
785
+ ARKitScenes_test Validation_48458650_7072.616_00000817
786
+ ARKitScenes_test Validation_48458650_7110.617_00003098
787
+ ARKitScenes_test Validation_48458656_7631.585_00000276
788
+ ARKitScenes_test Validation_48458656_7633.584_00000396
789
+ ARKitScenes_test Validation_48458656_7638.582_00000696
790
+ ARKitScenes_test Validation_48458656_7648.595_00001297
791
+ ARKitScenes_test Validation_48458656_7649.594_00001357
792
+ ARKitScenes_test Validation_48458657_7594.684_00000642
793
+ ARKitScenes_test Validation_48458657_7609.694_00001543
794
+ ARKitScenes_test Validation_48458660_8266.963_00001418
795
+ ARKitScenes_test Validation_48458667_8357.659_00000157
796
+ ARKitScenes_test Validation_48458667_8361.657_00000397
797
+ ARKitScenes_test Validation_48458667_8379.649_00001477
798
+ ARKitScenes_test Validation_48458667_8382.648_00001657
799
+ ARKitScenes_test Validation_48458667_8394.660_00002378
800
+ ARKitScenes_test Validation_48458667_8420.649_00003938
external/WildCamera/splits/arkitscenes_train.txt ADDED
The diff for this file is too large to render. See raw diff
 
external/WildCamera/splits/arkitscenes_val.txt ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ARKitScenes_test Validation_41069021_370.367_00003945
2
+ ARKitScenes_test Validation_41069021_387.360_00004965
3
+ ARKitScenes_test Validation_41069021_394.358_00005385
4
+ ARKitScenes_test Validation_41069021_406.369_00006106
5
+ ARKitScenes_test Validation_41069021_409.368_00006286
6
+ ARKitScenes_test Validation_41069021_425.362_00007246
7
+ ARKitScenes_test Validation_41069021_426.361_00007306
8
+ ARKitScenes_test Validation_41069021_430.359_00007546
9
+ ARKitScenes_test Validation_41069021_431.359_00007606
10
+ ARKitScenes_test Validation_41069021_444.370_00008387
11
+ ARKitScenes_test Validation_41069021_471.359_00010007
12
+ ARKitScenes_test Validation_41069021_488.369_00011028
13
+ ARKitScenes_test Validation_41069021_492.367_00011268
14
+ ARKitScenes_test Validation_41069025_607.870_00001117
15
+ ARKitScenes_test Validation_41069025_649.869_00003638
16
+ ARKitScenes_test Validation_41069025_701.864_00006759
17
+ ARKitScenes_test Validation_41069025_711.860_00007359
18
+ ARKitScenes_test Validation_41069025_717.858_00007719
19
+ ARKitScenes_test Validation_41069025_730.869_00008500
20
+ ARKitScenes_test Validation_41069025_735.867_00008800
21
+ ARKitScenes_test Validation_41069042_3049.237_00000385
22
+ ARKitScenes_test Validation_41069042_3069.228_00001585
23
+ ARKitScenes_test Validation_41069042_3071.227_00001705
24
+ ARKitScenes_test Validation_41069042_3105.230_00003746
25
+ ARKitScenes_test Validation_41069043_2860.631_00003698
26
+ ARKitScenes_test Validation_41069043_2893.634_00005679
27
+ ARKitScenes_test Validation_41069046_2921.323_00000636
28
+ ARKitScenes_test Validation_41069046_2923.339_00000757
29
+ ARKitScenes_test Validation_41069046_2925.338_00000877
30
+ ARKitScenes_test Validation_41069046_2930.336_00001177
31
+ ARKitScenes_test Validation_41069046_2960.323_00002977
32
+ ARKitScenes_test Validation_41069046_2963.322_00003157
33
+ ARKitScenes_test Validation_41069046_2980.332_00004178
34
+ ARKitScenes_test Validation_41142278_4002.081_00000037
35
+ ARKitScenes_test Validation_41142278_4042.081_00002438
36
+ ARKitScenes_test Validation_41142280_3836.982_00000877
37
+ ARKitScenes_test Validation_42444946_223633.870_00001597
38
+ ARKitScenes_test Validation_42444946_223675.869_00004118
39
+ ARKitScenes_test Validation_42444946_223682.866_00004538
40
+ ARKitScenes_test Validation_42444949_223780.677_00004479
41
+ ARKitScenes_test Validation_42444949_223783.675_00004659
42
+ ARKitScenes_test Validation_42444950_223809.565_00000697
43
+ ARKitScenes_test Validation_42444950_223855.563_00003458
44
+ ARKitScenes_test Validation_42444966_40343.793_00000883
45
+ ARKitScenes_test Validation_42444966_40380.794_00003104
46
+ ARKitScenes_test Validation_42444966_40384.793_00003344
47
+ ARKitScenes_test Validation_42444966_40393.789_00003884
48
+ ARKitScenes_test Validation_42444966_40402.785_00004424
49
+ ARKitScenes_test Validation_42444966_40406.783_00004664
50
+ ARKitScenes_test Validation_42444966_40407.783_00004724
51
+ ARKitScenes_test Validation_42444968_40245.400_00000049
52
+ ARKitScenes_test Validation_42444968_40246.383_00000108
53
+ ARKitScenes_test Validation_42444968_40252.381_00000468
54
+ ARKitScenes_test Validation_42444968_40258.395_00000829
55
+ ARKitScenes_test Validation_42444968_40297.396_00003170
56
+ ARKitScenes_test Validation_42444968_40300.394_00003350
57
+ ARKitScenes_test Validation_42444968_40312.389_00004070
58
+ ARKitScenes_test Validation_42444976_40143.393_00000157
59
+ ARKitScenes_test Validation_42444976_40169.382_00001717
60
+ ARKitScenes_test Validation_42444976_40170.382_00001777
61
+ ARKitScenes_test Validation_42444976_40171.398_00001838
62
+ ARKitScenes_test Validation_42444976_40182.393_00002498
63
+ ARKitScenes_test Validation_42444976_40186.392_00002738
64
+ ARKitScenes_test Validation_42444976_40198.387_00003458
65
+ ARKitScenes_test Validation_42444976_40199.386_00003518
66
+ ARKitScenes_test Validation_42445021_48884.258_00000276
67
+ ARKitScenes_test Validation_42445021_48887.257_00000456
68
+ ARKitScenes_test Validation_42445022_49008.276_00000049
69
+ ARKitScenes_test Validation_42445022_49014.256_00000408
70
+ ARKitScenes_test Validation_42445022_49015.256_00000468
71
+ ARKitScenes_test Validation_42445022_49035.265_00001669
72
+ ARKitScenes_test Validation_42445022_49070.268_00003770
73
+ ARKitScenes_test Validation_42445028_49607.697_00000349
74
+ ARKitScenes_test Validation_42445028_49609.696_00000469
75
+ ARKitScenes_test Validation_42445028_49655.695_00003230
76
+ ARKitScenes_test Validation_42445429_50330.084_00000817
77
+ ARKitScenes_test Validation_42445429_50333.083_00000997
78
+ ARKitScenes_test Validation_42445441_50300.796_00005196
79
+ ARKitScenes_test Validation_42445448_50410.586_00000115
80
+ ARKitScenes_test Validation_42445448_50486.589_00004677
81
+ ARKitScenes_test Validation_42445448_50494.586_00005157
82
+ ARKitScenes_test Validation_42445991_75264.421_00000522
83
+ ARKitScenes_test Validation_42445991_75304.421_00002923
84
+ ARKitScenes_test Validation_42445991_75326.429_00004244
85
+ ARKitScenes_test Validation_42446038_347909.974_00003044
86
+ ARKitScenes_test Validation_42446038_347925.967_00004004
87
+ ARKitScenes_test Validation_42446038_347928.983_00004185
88
+ ARKitScenes_test Validation_42446100_77913.295_00000397
89
+ ARKitScenes_test Validation_42446100_77962.292_00003338
90
+ ARKitScenes_test Validation_42446100_77966.290_00003578
91
+ ARKitScenes_test Validation_42446100_77994.296_00005259
92
+ ARKitScenes_test Validation_42446100_77997.295_00005439
93
+ ARKitScenes_test Validation_42446100_77999.294_00005559
94
+ ARKitScenes_test Validation_42446100_78006.291_00005979
95
+ ARKitScenes_test Validation_42446100_78007.291_00006039
96
+ ARKitScenes_test Validation_42446100_78017.287_00006639
97
+ ARKitScenes_test Validation_42446100_78019.286_00006759
98
+ ARKitScenes_test Validation_42446100_78027.283_00007239
99
+ ARKitScenes_test Validation_42446100_78028.282_00007299
100
+ ARKitScenes_test Validation_42446103_78062.186_00000583
101
+ ARKitScenes_test Validation_42446103_78083.194_00001844
102
+ ARKitScenes_test Validation_42446103_78090.191_00002264
103
+ ARKitScenes_test Validation_42446103_78096.189_00002624
104
+ ARKitScenes_test Validation_42446103_78130.192_00004665
105
+ ARKitScenes_test Validation_42446103_78137.189_00005085
106
+ ARKitScenes_test Validation_42446103_78139.188_00005205
107
+ ARKitScenes_test Validation_42446103_78157.181_00006285
108
+ ARKitScenes_test Validation_42446103_78177.190_00007486
109
+ ARKitScenes_test Validation_42446114_78198.981_00000637
110
+ ARKitScenes_test Validation_42446114_78214.992_00001598
111
+ ARKitScenes_test Validation_42446114_78222.989_00002078
112
+ ARKitScenes_test Validation_42446114_78245.996_00003459
113
+ ARKitScenes_test Validation_42446114_78251.994_00003819
114
+ ARKitScenes_test Validation_42446114_78272.986_00005079
115
+ ARKitScenes_test Validation_42446114_78292.994_00006280
116
+ ARKitScenes_test Validation_42446114_78297.992_00006580
117
+ ARKitScenes_test Validation_42446114_78298.992_00006640
118
+ ARKitScenes_test Validation_42446114_78306.989_00007120
119
+ ARKitScenes_test Validation_42446114_78311.987_00007420
120
+ ARKitScenes_test Validation_42446114_78324.982_00008200
121
+ ARKitScenes_test Validation_42446114_78367.981_00010781
122
+ ARKitScenes_test Validation_42446156_79137.122_00000097
123
+ ARKitScenes_test Validation_42446156_79149.117_00000817
124
+ ARKitScenes_test Validation_42446156_79224.121_00005319
125
+ ARKitScenes_test Validation_42446163_79569.523_00000217
126
+ ARKitScenes_test Validation_42446163_79570.522_00000277
127
+ ARKitScenes_test Validation_42446163_79603.526_00002258
128
+ ARKitScenes_test Validation_42446163_79604.525_00002318
129
+ ARKitScenes_test Validation_42446163_79608.524_00002558
130
+ ARKitScenes_test Validation_42446163_79619.519_00003218
131
+ ARKitScenes_test Validation_42446163_79627.533_00003699
132
+ ARKitScenes_test Validation_42446165_79718.730_00000217
133
+ ARKitScenes_test Validation_42446165_79724.728_00000577
134
+ ARKitScenes_test Validation_42446165_79745.719_00001837
135
+ ARKitScenes_test Validation_42446165_79771.726_00003398
136
+ ARKitScenes_test Validation_42446165_79792.734_00004659
137
+ ARKitScenes_test Validation_42446167_79644.726_00000103
138
+ ARKitScenes_test Validation_42446167_79645.726_00000163
139
+ ARKitScenes_test Validation_42446167_79654.722_00000703
140
+ ARKitScenes_test Validation_42446167_79676.730_00002024
141
+ ARKitScenes_test Validation_42446167_79682.728_00002384
142
+ ARKitScenes_test Validation_42446167_79707.734_00003885
143
+ ARKitScenes_test Validation_42446517_197447.270_00001777
144
+ ARKitScenes_test Validation_42446519_197506.878_00001358
145
+ ARKitScenes_test Validation_42446522_197316.873_00000186
146
+ ARKitScenes_test Validation_42446522_197359.872_00002767
147
+ ARKitScenes_test Validation_42446522_197360.872_00002827
148
+ ARKitScenes_test Validation_42446522_197384.879_00004268
149
+ ARKitScenes_test Validation_42446522_197393.875_00004808
150
+ ARKitScenes_test Validation_42446522_197396.874_00004988
151
+ ARKitScenes_test Validation_42446527_198633.848_00000481
152
+ ARKitScenes_test Validation_42446527_198646.843_00001261
153
+ ARKitScenes_test Validation_42446527_198667.851_00002522
154
+ ARKitScenes_test Validation_42446527_198696.839_00004262
155
+ ARKitScenes_test Validation_42446529_198739.838_00001093
156
+ ARKitScenes_test Validation_42446529_198779.838_00003494
157
+ ARKitScenes_test Validation_42446532_198859.939_00003140
158
+ ARKitScenes_test Validation_42446532_198900.938_00005601
159
+ ARKitScenes_test Validation_42446533_200400.295_00000415
160
+ ARKitScenes_test Validation_42446533_200426.301_00001976
161
+ ARKitScenes_test Validation_42446536_200466.901_00001375
162
+ ARKitScenes_test Validation_42446536_200467.901_00001435
163
+ ARKitScenes_test Validation_42446540_201470.345_00001171
164
+ ARKitScenes_test Validation_42446540_201486.338_00002131
165
+ ARKitScenes_test Validation_42446540_201487.338_00002191
166
+ ARKitScenes_test Validation_42446540_201494.335_00002611
167
+ ARKitScenes_test Validation_42446540_201498.350_00002852
168
+ ARKitScenes_test Validation_42446540_201513.344_00003752
169
+ ARKitScenes_test Validation_42446540_201518.342_00004052
170
+ ARKitScenes_test Validation_42446540_201530.337_00004772
171
+ ARKitScenes_test Validation_42446541_201561.441_00000823
172
+ ARKitScenes_test Validation_42446541_201570.437_00001363
173
+ ARKitScenes_test Validation_42446541_201594.444_00002804
174
+ ARKitScenes_test Validation_42446541_201622.449_00004485
175
+ ARKitScenes_test Validation_42446541_201644.440_00005805
176
+ ARKitScenes_test Validation_42446541_201645.439_00005865
177
+ ARKitScenes_test Validation_42446541_201665.448_00007066
178
+ ARKitScenes_test Validation_42897501_270038.590_00001826
179
+ ARKitScenes_test Validation_42897501_270046.587_00002306
180
+ ARKitScenes_test Validation_42897501_270054.584_00002786
181
+ ARKitScenes_test Validation_42897501_270057.583_00002966
182
+ ARKitScenes_test Validation_42897501_270063.580_00003326
183
+ ARKitScenes_test Validation_42897504_269851.484_00000559
184
+ ARKitScenes_test Validation_42897504_269871.492_00001760
185
+ ARKitScenes_test Validation_42897504_269876.490_00002060
186
+ ARKitScenes_test Validation_42897504_269882.488_00002420
187
+ ARKitScenes_test Validation_42897508_269924.104_00000104
188
+ ARKitScenes_test Validation_42897508_269929.085_00000403
189
+ ARKitScenes_test Validation_42897508_269938.082_00000943
190
+ ARKitScenes_test Validation_42897508_269951.093_00001724
191
+ ARKitScenes_test Validation_42897508_269954.092_00001904
192
+ ARKitScenes_test Validation_42897508_269959.090_00002204
193
+ ARKitScenes_test Validation_42897508_269964.088_00002504
194
+ ARKitScenes_test Validation_42897508_269971.085_00002924
195
+ ARKitScenes_test Validation_42897521_273955.896_00000181
196
+ ARKitScenes_test Validation_42897521_273961.894_00000541
197
+ ARKitScenes_test Validation_42897521_273987.900_00002102
198
+ ARKitScenes_test Validation_42897526_273849.407_00000295
199
+ ARKitScenes_test Validation_42897526_273883.393_00002335
200
+ ARKitScenes_test Validation_42897526_273887.391_00002575
201
+ ARKitScenes_test Validation_42897526_273889.390_00002695
202
+ ARKitScenes_test Validation_42897528_273912.098_00000721
203
+ ARKitScenes_test Validation_42897528_273918.095_00001081
204
+ ARKitScenes_test Validation_42897542_275467.131_00000751
205
+ ARKitScenes_test Validation_42897542_275474.128_00001171
206
+ ARKitScenes_test Validation_42897542_275482.125_00001651
207
+ ARKitScenes_test Validation_42897545_276443.861_00000342
208
+ ARKitScenes_test Validation_42897549_276516.364_00001867
209
+ ARKitScenes_test Validation_42897550_276550.167_00001057
210
+ ARKitScenes_test Validation_42897552_277375.808_00001057
211
+ ARKitScenes_test Validation_42897559_277439.598_00000541
212
+ ARKitScenes_test Validation_42897559_277473.601_00002582
213
+ ARKitScenes_test Validation_42897559_277476.600_00002762
214
+ ARKitScenes_test Validation_42897559_277477.599_00002822
215
+ ARKitScenes_test Validation_42897561_278935.714_00001459
216
+ ARKitScenes_test Validation_42897561_278941.712_00001819
217
+ ARKitScenes_test Validation_42897561_278944.711_00001999
218
+ ARKitScenes_test Validation_42897561_278957.705_00002779
219
+ ARKitScenes_test Validation_42897564_278991.308_00001237
220
+ ARKitScenes_test Validation_42897564_279023.311_00003158
221
+ ARKitScenes_test Validation_42897566_279059.013_00001196
222
+ ARKitScenes_test Validation_42897566_279087.001_00002876
223
+ ARKitScenes_test Validation_42897599_476337.198_00002149
224
+ ARKitScenes_test Validation_42897599_476424.212_00007372
225
+ ARKitScenes_test Validation_42897599_476451.201_00008992
226
+ ARKitScenes_test Validation_42897599_476465.212_00009833
227
+ ARKitScenes_test Validation_42897599_476473.209_00010313
228
+ ARKitScenes_test Validation_42897599_476510.210_00012534
229
+ ARKitScenes_test Validation_42897599_476511.210_00012594
230
+ ARKitScenes_test Validation_42897599_476548.211_00014815
231
+ ARKitScenes_test Validation_42897599_476553.209_00015115
232
+ ARKitScenes_test Validation_42897599_476562.206_00015655
233
+ ARKitScenes_test Validation_42897647_478138.188_00000523
234
+ ARKitScenes_test Validation_42897647_478149.183_00001183
235
+ ARKitScenes_test Validation_42897647_478150.183_00001243
236
+ ARKitScenes_test Validation_42897647_478160.179_00001843
237
+ ARKitScenes_test Validation_42897647_478167.176_00002263
238
+ ARKitScenes_test Validation_42897647_478170.175_00002443
239
+ ARKitScenes_test Validation_42897647_478208.176_00004724
240
+ ARKitScenes_test Validation_42897647_478226.185_00005805
241
+ ARKitScenes_test Validation_42897651_478328.276_00005036
242
+ ARKitScenes_test Validation_42897651_478335.274_00005456
243
+ ARKitScenes_test Validation_42897651_478342.287_00005877
244
+ ARKitScenes_test Validation_42897651_478353.283_00006537
245
+ ARKitScenes_test Validation_42897651_478356.282_00006717
246
+ ARKitScenes_test Validation_42897651_478443.279_00011939
247
+ ARKitScenes_test Validation_42897651_478480.280_00014160
248
+ ARKitScenes_test Validation_42897651_478487.278_00014580
249
+ ARKitScenes_test Validation_42897651_478519.281_00016501
250
+ ARKitScenes_test Validation_42897651_478543.271_00017941
251
+ ARKitScenes_test Validation_42897667_478004.076_00000456
252
+ ARKitScenes_test Validation_42897667_478015.088_00001117
253
+ ARKitScenes_test Validation_42897667_478031.082_00002077
254
+ ARKitScenes_test Validation_42897667_478032.081_00002137
255
+ ARKitScenes_test Validation_42897667_478033.081_00002197
256
+ ARKitScenes_test Validation_42897667_478053.089_00003398
257
+ ARKitScenes_test Validation_42897667_478083.077_00005198
258
+ ARKitScenes_test Validation_42897667_478117.080_00007239
259
+ ARKitScenes_test Validation_42897688_479183.081_00002377
260
+ ARKitScenes_test Validation_42897688_479204.090_00003638
261
+ ARKitScenes_test Validation_42897688_479223.082_00004778
262
+ ARKitScenes_test Validation_42897688_479228.080_00005078
263
+ ARKitScenes_test Validation_42897688_479253.086_00006579
264
+ ARKitScenes_test Validation_42897688_479277.076_00008019
265
+ ARKitScenes_test Validation_42897688_479292.087_00008920
266
+ ARKitScenes_test Validation_42898811_167510.387_00002377
267
+ ARKitScenes_test Validation_42898816_167394.902_00002504
268
+ ARKitScenes_test Validation_42898818_168161.177_00000577
269
+ ARKitScenes_test Validation_42898818_168172.172_00001237
270
+ ARKitScenes_test Validation_42898818_168173.172_00001297
271
+ ARKitScenes_test Validation_42898822_168209.574_00000277
272
+ ARKitScenes_test Validation_42898822_168250.573_00002738
273
+ ARKitScenes_test Validation_42898826_168118.178_00001243
274
+ ARKitScenes_test Validation_42898826_168119.178_00001303
275
+ ARKitScenes_test Validation_42898849_196459.913_00000877
276
+ ARKitScenes_test Validation_42898849_196460.913_00000937
277
+ ARKitScenes_test Validation_42898849_196468.926_00001418
278
+ ARKitScenes_test Validation_42898849_196471.925_00001598
279
+ ARKitScenes_test Validation_42898849_196503.912_00003518
280
+ ARKitScenes_test Validation_42898849_196531.918_00005199
281
+ ARKitScenes_test Validation_42898849_196542.913_00005859
282
+ ARKitScenes_test Validation_42898849_196550.927_00006340
283
+ ARKitScenes_test Validation_42898849_196566.921_00007300
284
+ ARKitScenes_test Validation_42898849_196584.913_00008380
285
+ ARKitScenes_test Validation_42898854_196220.724_00002288
286
+ ARKitScenes_test Validation_42898854_196229.721_00002828
287
+ ARKitScenes_test Validation_42898854_196233.719_00003068
288
+ ARKitScenes_test Validation_42898854_196236.718_00003248
289
+ ARKitScenes_test Validation_42898854_196266.723_00005049
290
+ ARKitScenes_test Validation_42898854_196272.720_00005409
291
+ ARKitScenes_test Validation_42898854_196274.720_00005529
292
+ ARKitScenes_test Validation_42898854_196285.715_00006189
293
+ ARKitScenes_test Validation_42898862_196315.720_00000271
294
+ ARKitScenes_test Validation_42898862_196349.723_00002312
295
+ ARKitScenes_test Validation_42898862_196366.716_00003332
296
+ ARKitScenes_test Validation_42898862_196372.714_00003692
297
+ ARKitScenes_test Validation_42898862_196375.713_00003872
298
+ ARKitScenes_test Validation_42898862_196379.711_00004112
299
+ ARKitScenes_test Validation_42898862_196388.724_00004653
300
+ ARKitScenes_test Validation_42898862_196400.720_00005373
301
+ ARKitScenes_test Validation_42898862_196401.719_00005433
302
+ ARKitScenes_test Validation_42898862_196404.718_00005613
303
+ ARKitScenes_test Validation_42898862_196405.718_00005673
304
+ ARKitScenes_test Validation_42898862_196416.713_00006333
305
+ ARKitScenes_test Validation_42898862_196426.726_00006934
306
+ ARKitScenes_test Validation_42899461_195789.434_00000277
307
+ ARKitScenes_test Validation_42899461_195799.430_00000877
308
+ ARKitScenes_test Validation_42899461_195831.433_00002798
309
+ ARKitScenes_test Validation_42899461_195850.425_00003938
310
+ ARKitScenes_test Validation_42899461_195859.421_00004478
311
+ ARKitScenes_test Validation_42899461_195866.418_00004898
312
+ ARKitScenes_test Validation_42899471_195717.530_00003579
313
+ ARKitScenes_test Validation_42899471_195718.530_00003639
314
+ ARKitScenes_test Validation_42899611_228351.719_00000157
315
+ ARKitScenes_test Validation_42899611_228363.714_00000877
316
+ ARKitScenes_test Validation_42899611_228389.720_00002438
317
+ ARKitScenes_test Validation_42899611_228392.719_00002618
318
+ ARKitScenes_test Validation_42899611_228395.718_00002798
319
+ ARKitScenes_test Validation_42899612_228436.519_00000649
320
+ ARKitScenes_test Validation_42899612_228440.517_00000889
321
+ ARKitScenes_test Validation_42899612_228494.512_00004130
322
+ ARKitScenes_test Validation_42899612_228495.512_00004190
323
+ ARKitScenes_test Validation_42899612_228504.525_00004731
324
+ ARKitScenes_test Validation_42899612_228518.520_00005571
325
+ ARKitScenes_test Validation_42899613_228276.016_00000222
326
+ ARKitScenes_test Validation_42899613_228281.014_00000522
327
+ ARKitScenes_test Validation_42899613_228313.018_00002443
328
+ ARKitScenes_test Validation_42899613_228327.012_00003283
329
+ ARKitScenes_test Validation_42899619_228891.872_00000217
330
+ ARKitScenes_test Validation_42899619_228897.869_00000577
331
+ ARKitScenes_test Validation_42899620_228853.670_00000397
332
+ ARKitScenes_test Validation_42899620_228857.669_00000637
333
+ ARKitScenes_test Validation_44358442_42920.105_00001784
334
+ ARKitScenes_test Validation_44358446_43014.899_00000397
335
+ ARKitScenes_test Validation_44358446_43015.898_00000457
336
+ ARKitScenes_test Validation_44358446_43085.903_00004659
337
+ ARKitScenes_test Validation_44358446_43089.901_00004899
338
+ ARKitScenes_test Validation_44358446_43101.896_00005619
339
+ ARKitScenes_test Validation_44358446_43119.889_00006699
340
+ ARKitScenes_test Validation_44358448_42964.803_00000895
341
+ ARKitScenes_test Validation_44358451_43632.897_00001658
342
+ ARKitScenes_test Validation_44358451_43712.880_00006459
343
+ ARKitScenes_test Validation_44358452_43503.584_00000216
344
+ ARKitScenes_test Validation_44358452_43538.586_00002317
345
+ ARKitScenes_test Validation_44358455_43441.293_00002858
346
+ ARKitScenes_test Validation_44358455_43481.293_00005259
347
+ ARKitScenes_test Validation_45260854_561.091_00000289
348
+ ARKitScenes_test Validation_45260854_568.088_00000709
349
+ ARKitScenes_test Validation_45260854_570.087_00000829
350
+ ARKitScenes_test Validation_45260854_582.082_00001549
351
+ ARKitScenes_test Validation_45260856_640.292_00000877
352
+ ARKitScenes_test Validation_45260856_648.288_00001357
353
+ ARKitScenes_test Validation_45260856_653.286_00001657
354
+ ARKitScenes_test Validation_45260856_698.284_00004358
355
+ ARKitScenes_test Validation_45260857_519.692_00001358
356
+ ARKitScenes_test Validation_45260857_529.687_00001958
357
+ ARKitScenes_test Validation_45260857_534.685_00002258
358
+ ARKitScenes_test Validation_45260857_536.685_00002378
359
+ ARKitScenes_test Validation_45260857_539.683_00002558
360
+ ARKitScenes_test Validation_45260898_769.990_00000409
361
+ ARKitScenes_test Validation_45260898_774.988_00000709
362
+ ARKitScenes_test Validation_45260899_868.784_00001243
363
+ ARKitScenes_test Validation_45260899_885.794_00002264
364
+ ARKitScenes_test Validation_45260899_895.790_00002864
365
+ ARKitScenes_test Validation_45260900_804.493_00000073
366
+ ARKitScenes_test Validation_45260920_2471.417_00000288
367
+ ARKitScenes_test Validation_45260920_2474.416_00000468
368
+ ARKitScenes_test Validation_45260920_2494.425_00001669
369
+ ARKitScenes_test Validation_45260920_2512.417_00002749
370
+ ARKitScenes_test Validation_45260920_2535.425_00004130
371
+ ARKitScenes_test Validation_45260925_2412.823_00001117
372
+ ARKitScenes_test Validation_45260925_2424.819_00001837
373
+ ARKitScenes_test Validation_45260925_2427.818_00002017
374
+ ARKitScenes_test Validation_45260925_2454.824_00003638
375
+ ARKitScenes_test Validation_45260928_2345.217_00000457
376
+ ARKitScenes_test Validation_45260928_2362.227_00001478
377
+ ARKitScenes_test Validation_45260928_2371.223_00002018
378
+ ARKitScenes_test Validation_45261121_3447.063_00001994
379
+ ARKitScenes_test Validation_45261121_3448.063_00002054
380
+ ARKitScenes_test Validation_45261121_3454.060_00002414
381
+ ARKitScenes_test Validation_45261121_3468.071_00003255
382
+ ARKitScenes_test Validation_45261128_3526.164_00002198
383
+ ARKitScenes_test Validation_45261128_3536.159_00002798
384
+ ARKitScenes_test Validation_45261133_4030.272_00001417
385
+ ARKitScenes_test Validation_45261133_4040.268_00002017
386
+ ARKitScenes_test Validation_45261133_4047.265_00002437
387
+ ARKitScenes_test Validation_45261133_4048.265_00002497
388
+ ARKitScenes_test Validation_45261133_4050.264_00002617
389
+ ARKitScenes_test Validation_45261133_4079.268_00004358
390
+ ARKitScenes_test Validation_45261133_4080.268_00004418
391
+ ARKitScenes_test Validation_45261133_4107.273_00006039
392
+ ARKitScenes_test Validation_45261133_4116.270_00006579
393
+ ARKitScenes_test Validation_45261133_4117.269_00006639
394
+ ARKitScenes_test Validation_45261133_4126.266_00007179
395
+ ARKitScenes_test Validation_45261133_4132.263_00007539
396
+ ARKitScenes_test Validation_45261133_4136.261_00007779
397
+ ARKitScenes_test Validation_45261133_4142.259_00008139
398
+ ARKitScenes_test Validation_45261133_4148.256_00008499
399
+ ARKitScenes_test Validation_45261140_3951.272_00000283
400
+ ARKitScenes_test Validation_45261140_3960.268_00000823
401
+ ARKitScenes_test Validation_45261140_3976.261_00001783
402
+ ARKitScenes_test Validation_45261140_3978.260_00001903
403
+ ARKitScenes_test Validation_45261142_3888.464_00000174
404
+ ARKitScenes_test Validation_45261142_3896.461_00000654
405
+ ARKitScenes_test Validation_45261143_4588.703_00000583
406
+ ARKitScenes_test Validation_45261143_4592.701_00000823
407
+ ARKitScenes_test Validation_45261143_4628.703_00002984
408
+ ARKitScenes_test Validation_45261144_4658.007_00000457
409
+ ARKitScenes_test Validation_45261150_4531.310_00000337
410
+ ARKitScenes_test Validation_45261150_4535.308_00000577
411
+ ARKitScenes_test Validation_45261150_4549.302_00001417
412
+ ARKitScenes_test Validation_45261181_2852.026_00001537
413
+ ARKitScenes_test Validation_45261182_2977.726_00003818
414
+ ARKitScenes_test Validation_45261182_2994.736_00004839
415
+ ARKitScenes_test Validation_45261185_3558.482_00000703
416
+ ARKitScenes_test Validation_45261185_3559.481_00000763
417
+ ARKitScenes_test Validation_45261185_3562.480_00000943
418
+ ARKitScenes_test Validation_45261185_3601.481_00003284
419
+ ARKitScenes_test Validation_45261190_3716.886_00001843
420
+ ARKitScenes_test Validation_45261190_3743.892_00003464
421
+ ARKitScenes_test Validation_45261193_3625.189_00000409
422
+ ARKitScenes_test Validation_45261193_3632.186_00000829
423
+ ARKitScenes_test Validation_45261193_3655.193_00002210
424
+ ARKitScenes_test Validation_45261575_77474.508_00000337
425
+ ARKitScenes_test Validation_45261575_77475.507_00000397
426
+ ARKitScenes_test Validation_45261581_77518.506_00000157
427
+ ARKitScenes_test Validation_45261581_77535.516_00001178
428
+ ARKitScenes_test Validation_45261594_78039.207_00000637
429
+ ARKitScenes_test Validation_45261615_55008.692_00000703
430
+ ARKitScenes_test Validation_45261619_55088.693_00001436
431
+ ARKitScenes_test Validation_45261619_55116.682_00003116
432
+ ARKitScenes_test Validation_45261620_54934.288_00000157
433
+ ARKitScenes_test Validation_45261631_56030.404_00000217
434
+ ARKitScenes_test Validation_45261631_56068.389_00002497
435
+ ARKitScenes_test Validation_45261631_56089.397_00003758
436
+ ARKitScenes_test Validation_45261631_56092.396_00003938
437
+ ARKitScenes_test Validation_45261631_56104.391_00004658
438
+ ARKitScenes_test Validation_45261632_55892.192_00000138
439
+ ARKitScenes_test Validation_45261637_55961.498_00000337
440
+ ARKitScenes_test Validation_45261637_55962.497_00000397
441
+ ARKitScenes_test Validation_45261637_55971.494_00000937
442
+ ARKitScenes_test Validation_45662921_21784.929_00002048
443
+ ARKitScenes_test Validation_45662921_21787.928_00002228
444
+ ARKitScenes_test Validation_45662921_21791.926_00002468
445
+ ARKitScenes_test Validation_45662921_21836.924_00005169
446
+ ARKitScenes_test Validation_45662924_21592.925_00002258
447
+ ARKitScenes_test Validation_45662924_21607.919_00003158
448
+ ARKitScenes_test Validation_45662924_21609.918_00003278
449
+ ARKitScenes_test Validation_45662926_21649.319_00000667
450
+ ARKitScenes_test Validation_45662926_21659.331_00001268
451
+ ARKitScenes_test Validation_45662926_21670.326_00001928
452
+ ARKitScenes_test Validation_45662926_21674.325_00002168
453
+ ARKitScenes_test Validation_45662926_21675.324_00002228
454
+ ARKitScenes_test Validation_45662926_21682.322_00002648
455
+ ARKitScenes_test Validation_45662926_21684.321_00002768
456
+ ARKitScenes_test Validation_45662926_21688.319_00003008
457
+ ARKitScenes_test Validation_45662926_21692.317_00003248
458
+ ARKitScenes_test Validation_45662926_21693.317_00003308
459
+ ARKitScenes_test Validation_45662926_21694.317_00003368
460
+ ARKitScenes_test Validation_45662926_21711.326_00004389
461
+ ARKitScenes_test Validation_45662926_21715.324_00004629
462
+ ARKitScenes_test Validation_45662942_22918.660_00002197
463
+ ARKitScenes_test Validation_45662943_23032.263_00000817
464
+ ARKitScenes_test Validation_45662943_23164.258_00008740
465
+ ARKitScenes_test Validation_45662943_23168.256_00008980
466
+ ARKitScenes_test Validation_45662943_23170.255_00009100
467
+ ARKitScenes_test Validation_45662943_23171.255_00009160
468
+ ARKitScenes_test Validation_45662943_23174.254_00009340
469
+ ARKitScenes_test Validation_45662943_23175.253_00009400
470
+ ARKitScenes_test Validation_45662943_23179.268_00009641
471
+ ARKitScenes_test Validation_45662943_23183.267_00009881
472
+ ARKitScenes_test Validation_45662943_23208.256_00011381
473
+ ARKitScenes_test Validation_45662944_22963.758_00001297
474
+ ARKitScenes_test Validation_45662944_22974.770_00001958
475
+ ARKitScenes_test Validation_45662970_26159.531_00003188
476
+ ARKitScenes_test Validation_45662979_26048.725_00001177
477
+ ARKitScenes_test Validation_45662981_27937.116_00000050
478
+ ARKitScenes_test Validation_45662981_27946.096_00000589
479
+ ARKitScenes_test Validation_45662981_27994.094_00003470
480
+ ARKitScenes_test Validation_45662981_28029.097_00005571
481
+ ARKitScenes_test Validation_45662983_27801.886_00000337
482
+ ARKitScenes_test Validation_45662983_27805.885_00000577
483
+ ARKitScenes_test Validation_45662983_27807.884_00000697
484
+ ARKitScenes_test Validation_45662983_27851.883_00003338
485
+ ARKitScenes_test Validation_45662983_27854.882_00003518
486
+ ARKitScenes_test Validation_45662987_27869.193_00000169
487
+ ARKitScenes_test Validation_45662987_27878.190_00000709
488
+ ARKitScenes_test Validation_45662987_27883.188_00001009
489
+ ARKitScenes_test Validation_45662987_27910.194_00002630
490
+ ARKitScenes_test Validation_45662987_27911.193_00002690
491
+ ARKitScenes_test Validation_45663113_55308.661_00000277
492
+ ARKitScenes_test Validation_45663113_55324.671_00001238
493
+ ARKitScenes_test Validation_45663113_55326.671_00001358
494
+ ARKitScenes_test Validation_45663114_55386.862_00000637
495
+ ARKitScenes_test Validation_45663114_55404.855_00001717
496
+ ARKitScenes_test Validation_45663115_55461.865_00001958
497
+ ARKitScenes_test Validation_45663149_63304.709_00000517
498
+ ARKitScenes_test Validation_45663149_63311.706_00000937
499
+ ARKitScenes_test Validation_45663149_63314.705_00001117
500
+ ARKitScenes_test Validation_45663149_63315.705_00001177
501
+ ARKitScenes_test Validation_45663149_63324.701_00001717
502
+ ARKitScenes_test Validation_45663150_63511.952_00000288
503
+ ARKitScenes_test Validation_45663150_63551.953_00002689
504
+ ARKitScenes_test Validation_45663154_63399.880_00000074
505
+ ARKitScenes_test Validation_45663154_63401.862_00000193
506
+ ARKitScenes_test Validation_45663154_63409.859_00000673
507
+ ARKitScenes_test Validation_45663154_63411.858_00000793
508
+ ARKitScenes_test Validation_45663154_63450.860_00003134
509
+ ARKitScenes_test Validation_45663164_65947.828_00001015
510
+ ARKitScenes_test Validation_45663164_65948.828_00001075
511
+ ARKitScenes_test Validation_45663164_65968.820_00002275
512
+ ARKitScenes_test Validation_45663164_65975.817_00002695
513
+ ARKitScenes_test Validation_45663164_65997.825_00004016
514
+ ARKitScenes_test Validation_45663165_65815.014_00000637
515
+ ARKitScenes_test Validation_45663175_65890.618_00001153
516
+ ARKitScenes_test Validation_45663175_65891.617_00001213
517
+ ARKitScenes_test Validation_45663175_65912.625_00002474
518
+ ARKitScenes_test Validation_45663175_65919.623_00002894
519
+ ARKitScenes_test Validation_47115452_13057.568_00000097
520
+ ARKitScenes_test Validation_47115452_13076.576_00001238
521
+ ARKitScenes_test Validation_47115460_13145.781_00000529
522
+ ARKitScenes_test Validation_47115460_13166.772_00001789
523
+ ARKitScenes_test Validation_47115463_13220.467_00000096
524
+ ARKitScenes_test Validation_47115463_13235.477_00000997
525
+ ARKitScenes_test Validation_47115463_13282.474_00003818
526
+ ARKitScenes_test Validation_47115469_14055.292_00002678
527
+ ARKitScenes_test Validation_47115469_14059.290_00002918
528
+ ARKitScenes_test Validation_47115473_14107.787_00002150
529
+ ARKitScenes_test Validation_47115473_14172.776_00006051
530
+ ARKitScenes_test Validation_47115474_13966.678_00000402
531
+ ARKitScenes_test Validation_47115474_13976.691_00001003
532
+ ARKitScenes_test Validation_47204552_3357.092_00000277
533
+ ARKitScenes_test Validation_47204552_3378.100_00001538
534
+ ARKitScenes_test Validation_47204552_3391.095_00002318
535
+ ARKitScenes_test Validation_47204552_3412.086_00003578
536
+ ARKitScenes_test Validation_47204552_3425.097_00004359
537
+ ARKitScenes_test Validation_47204552_3430.095_00004659
538
+ ARKitScenes_test Validation_47204552_3441.091_00005319
539
+ ARKitScenes_test Validation_47204552_3485.089_00007960
540
+ ARKitScenes_test Validation_47204552_3486.089_00008020
541
+ ARKitScenes_test Validation_47204554_3215.201_00000277
542
+ ARKitScenes_test Validation_47204554_3227.196_00000997
543
+ ARKitScenes_test Validation_47204554_3240.191_00001777
544
+ ARKitScenes_test Validation_47204554_3244.189_00002017
545
+ ARKitScenes_test Validation_47204554_3261.199_00003038
546
+ ARKitScenes_test Validation_47204554_3263.198_00003158
547
+ ARKitScenes_test Validation_47204554_3265.197_00003278
548
+ ARKitScenes_test Validation_47204554_3269.195_00003518
549
+ ARKitScenes_test Validation_47204556_3286.405_00000037
550
+ ARKitScenes_test Validation_47204556_3312.394_00001597
551
+ ARKitScenes_test Validation_47204559_3992.536_00000276
552
+ ARKitScenes_test Validation_47204559_4001.549_00000817
553
+ ARKitScenes_test Validation_47204563_4044.448_00000697
554
+ ARKitScenes_test Validation_47204566_4109.038_00002017
555
+ ARKitScenes_test Validation_47204566_4118.051_00002558
556
+ ARKitScenes_test Validation_47204573_5014.617_00003459
557
+ ARKitScenes_test Validation_47204573_5039.606_00004959
558
+ ARKitScenes_test Validation_47204573_5046.620_00005380
559
+ ARKitScenes_test Validation_47204578_4918.223_00001484
560
+ ARKitScenes_test Validation_47204605_2434.303_00001922
561
+ ARKitScenes_test Validation_47204607_2521.802_00000097
562
+ ARKitScenes_test Validation_47204607_2522.802_00000157
563
+ ARKitScenes_test Validation_47204607_2523.801_00000217
564
+ ARKitScenes_test Validation_47204607_2554.806_00002078
565
+ ARKitScenes_test Validation_47204607_2590.791_00004238
566
+ ARKitScenes_test Validation_47204607_2635.790_00006939
567
+ ARKitScenes_test Validation_47204607_2676.791_00009400
568
+ ARKitScenes_test Validation_47204607_2677.790_00009460
569
+ ARKitScenes_test Validation_47204609_2459.993_00000457
570
+ ARKitScenes_test Validation_47204609_2462.992_00000637
571
+ ARKitScenes_test Validation_47331068_131079.065_00000877
572
+ ARKitScenes_test Validation_47331071_131004.361_00000637
573
+ ARKitScenes_test Validation_47331265_90018.298_00000433
574
+ ARKitScenes_test Validation_47331265_90026.295_00000913
575
+ ARKitScenes_test Validation_47331266_89826.492_00000319
576
+ ARKitScenes_test Validation_47331266_89870.491_00002960
577
+ ARKitScenes_test Validation_47331266_89893.499_00004341
578
+ ARKitScenes_test Validation_47331319_179047.025_00000654
579
+ ARKitScenes_test Validation_47331322_179161.329_00002870
580
+ ARKitScenes_test Validation_47331337_183502.425_00001897
581
+ ARKitScenes_test Validation_47331339_183421.825_00001579
582
+ ARKitScenes_test Validation_47331339_183426.823_00001879
583
+ ARKitScenes_test Validation_47331651_22884.238_00002078
584
+ ARKitScenes_test Validation_47331651_22902.231_00003158
585
+ ARKitScenes_test Validation_47331651_22903.230_00003218
586
+ ARKitScenes_test Validation_47331651_22905.230_00003338
587
+ ARKitScenes_test Validation_47331651_22921.223_00004298
588
+ ARKitScenes_test Validation_47331651_22924.239_00004479
589
+ ARKitScenes_test Validation_47331651_22928.237_00004719
590
+ ARKitScenes_test Validation_47331651_22930.236_00004839
591
+ ARKitScenes_test Validation_47331651_22939.232_00005379
592
+ ARKitScenes_test Validation_47331651_22947.229_00005859
593
+ ARKitScenes_test Validation_47331651_22971.236_00007300
594
+ ARKitScenes_test Validation_47331651_22972.236_00007360
595
+ ARKitScenes_test Validation_47331653_22994.327_00000577
596
+ ARKitScenes_test Validation_47331653_23001.324_00000997
597
+ ARKitScenes_test Validation_47331653_23054.335_00004179
598
+ ARKitScenes_test Validation_47331653_23064.331_00004779
599
+ ARKitScenes_test Validation_47331653_23111.329_00007600
600
+ ARKitScenes_test Validation_47331653_23114.328_00007780
601
+ ARKitScenes_test Validation_47331653_23127.322_00008560
602
+ ARKitScenes_test Validation_47331653_23133.336_00008921
603
+ ARKitScenes_test Validation_47331653_23134.336_00008981
604
+ ARKitScenes_test Validation_47331654_22743.829_00000457
605
+ ARKitScenes_test Validation_47331654_22785.828_00002978
606
+ ARKitScenes_test Validation_47331654_22786.828_00003038
607
+ ARKitScenes_test Validation_47331654_22832.826_00005799
608
+ ARKitScenes_test Validation_47331661_23633.580_00002870
609
+ ARKitScenes_test Validation_47331661_23643.593_00003471
610
+ ARKitScenes_test Validation_47331661_23644.592_00003531
611
+ ARKitScenes_test Validation_47331661_23685.592_00005992
612
+ ARKitScenes_test Validation_47331661_23688.591_00006172
613
+ ARKitScenes_test Validation_47331661_23699.587_00006832
614
+ ARKitScenes_test Validation_47331661_23720.578_00008092
615
+ ARKitScenes_test Validation_47331662_23462.083_00002497
616
+ ARKitScenes_test Validation_47331662_23468.081_00002857
617
+ ARKitScenes_test Validation_47331662_23469.080_00002917
618
+ ARKitScenes_test Validation_47331662_23485.091_00003878
619
+ ARKitScenes_test Validation_47331662_23488.089_00004058
620
+ ARKitScenes_test Validation_47331988_170632.558_00002264
621
+ ARKitScenes_test Validation_47331988_170653.550_00003524
622
+ ARKitScenes_test Validation_47331988_170659.548_00003884
623
+ ARKitScenes_test Validation_47331988_170749.561_00009287
624
+ ARKitScenes_test Validation_47331989_170421.860_00000457
625
+ ARKitScenes_test Validation_47331989_170455.863_00002498
626
+ ARKitScenes_test Validation_47331990_170521.953_00002509
627
+ ARKitScenes_test Validation_47331990_170525.952_00002749
628
+ ARKitScenes_test Validation_47331990_170549.959_00004190
629
+ ARKitScenes_test Validation_47331990_170558.955_00004730
630
+ ARKitScenes_test Validation_47332005_172103.731_00002443
631
+ ARKitScenes_test Validation_47332005_172170.737_00006465
632
+ ARKitScenes_test Validation_47332885_7493.862_00000583
633
+ ARKitScenes_test Validation_47332886_7578.261_00002215
634
+ ARKitScenes_test Validation_47332890_7645.367_00002918
635
+ ARKitScenes_test Validation_47332890_7647.366_00003038
636
+ ARKitScenes_test Validation_47332890_7669.357_00004358
637
+ ARKitScenes_test Validation_47332895_9056.792_00000655
638
+ ARKitScenes_test Validation_47332899_9178.193_00003548
639
+ ARKitScenes_test Validation_47332901_11035.593_00001165
640
+ ARKitScenes_test Validation_47332901_11041.590_00001525
641
+ ARKitScenes_test Validation_47332901_11042.590_00001585
642
+ ARKitScenes_test Validation_47332904_11139.400_00003705
643
+ ARKitScenes_test Validation_47332904_11147.397_00004185
644
+ ARKitScenes_test Validation_47332905_10983.897_00002030
645
+ ARKitScenes_test Validation_47332910_12199.324_00000768
646
+ ARKitScenes_test Validation_47332910_12253.336_00004010
647
+ ARKitScenes_test Validation_47332911_12284.823_00001543
648
+ ARKitScenes_test Validation_47332911_12350.829_00005505
649
+ ARKitScenes_test Validation_47332915_13471.971_00000343
650
+ ARKitScenes_test Validation_47332915_13507.973_00002504
651
+ ARKitScenes_test Validation_47332915_13543.975_00004665
652
+ ARKitScenes_test Validation_47332916_13398.568_00000475
653
+ ARKitScenes_test Validation_47332918_13325.864_00000589
654
+ ARKitScenes_test Validation_47332918_13336.876_00001250
655
+ ARKitScenes_test Validation_47332918_13368.863_00003170
656
+ ARKitScenes_test Validation_47333440_53804.226_00000637
657
+ ARKitScenes_test Validation_47333440_53839.228_00002738
658
+ ARKitScenes_test Validation_47333441_53869.716_00001117
659
+ ARKitScenes_test Validation_47333443_53783.118_00002077
660
+ ARKitScenes_test Validation_47333443_53784.117_00002137
661
+ ARKitScenes_test Validation_47333452_54954.122_00000157
662
+ ARKitScenes_test Validation_47333452_54987.208_00002143
663
+ ARKitScenes_test Validation_47333456_55036.122_00001718
664
+ ARKitScenes_test Validation_47333898_81373.971_00003278
665
+ ARKitScenes_test Validation_47333899_81292.071_00003098
666
+ ARKitScenes_test Validation_47333904_81413.372_00000697
667
+ ARKitScenes_test Validation_47333904_81420.369_00001117
668
+ ARKitScenes_test Validation_47333904_81424.367_00001357
669
+ ARKitScenes_test Validation_47333904_81426.367_00001477
670
+ ARKitScenes_test Validation_47333904_81435.363_00002017
671
+ ARKitScenes_test Validation_47333904_81443.376_00002498
672
+ ARKitScenes_test Validation_47333904_81444.376_00002558
673
+ ARKitScenes_test Validation_47333916_82429.566_00003699
674
+ ARKitScenes_test Validation_47333916_82433.564_00003939
675
+ ARKitScenes_test Validation_47333918_82508.450_00003163
676
+ ARKitScenes_test Validation_47333918_82513.465_00003464
677
+ ARKitScenes_test Validation_47333920_82646.261_00006096
678
+ ARKitScenes_test Validation_47333920_82649.260_00006276
679
+ ARKitScenes_test Validation_47333920_82658.256_00006816
680
+ ARKitScenes_test Validation_47333920_82663.254_00007116
681
+ ARKitScenes_test Validation_47333923_83608.499_00000636
682
+ ARKitScenes_test Validation_47333923_83645.501_00002857
683
+ ARKitScenes_test Validation_47333924_83693.915_00002078
684
+ ARKitScenes_test Validation_47333925_83549.606_00000523
685
+ ARKitScenes_test Validation_47333925_83550.606_00000583
686
+ ARKitScenes_test Validation_47333925_83581.610_00002444
687
+ ARKitScenes_test Validation_47333927_84003.213_00000824
688
+ ARKitScenes_test Validation_47333927_84012.209_00001364
689
+ ARKitScenes_test Validation_47333931_83942.704_00001778
690
+ ARKitScenes_test Validation_47333932_84111.202_00002918
691
+ ARKitScenes_test Validation_47333932_84112.202_00002978
692
+ ARKitScenes_test Validation_47333932_84125.196_00003758
693
+ ARKitScenes_test Validation_47333932_84183.206_00007240
694
+ ARKitScenes_test Validation_47333934_85569.766_00001249
695
+ ARKitScenes_test Validation_47333934_85621.762_00004370
696
+ ARKitScenes_test Validation_47333934_85645.769_00005811
697
+ ARKitScenes_test Validation_47333934_85655.765_00006411
698
+ ARKitScenes_test Validation_47333934_85660.763_00006711
699
+ ARKitScenes_test Validation_47333937_85506.376_00001021
700
+ ARKitScenes_test Validation_47333937_85512.373_00001381
701
+ ARKitScenes_test Validation_47333940_85460.977_00003038
702
+ ARKitScenes_test Validation_47333940_85482.968_00004358
703
+ ARKitScenes_test Validation_47334105_92895.748_00000769
704
+ ARKitScenes_test Validation_47334105_92897.747_00000889
705
+ ARKitScenes_test Validation_47334105_92903.744_00001249
706
+ ARKitScenes_test Validation_47334105_92922.753_00002390
707
+ ARKitScenes_test Validation_47334105_92946.744_00003830
708
+ ARKitScenes_test Validation_47334105_92951.758_00004131
709
+ ARKitScenes_test Validation_47334105_92954.757_00004311
710
+ ARKitScenes_test Validation_47334105_92986.744_00006231
711
+ ARKitScenes_test Validation_47334105_92990.759_00006472
712
+ ARKitScenes_test Validation_47334105_93022.746_00008392
713
+ ARKitScenes_test Validation_47334105_93029.743_00008812
714
+ ARKitScenes_test Validation_47334239_111112.020_00000517
715
+ ARKitScenes_test Validation_47334239_111149.022_00002738
716
+ ARKitScenes_test Validation_47334240_111193.421_00002017
717
+ ARKitScenes_test Validation_47334240_111214.429_00003278
718
+ ARKitScenes_test Validation_47334256_112032.695_00002258
719
+ ARKitScenes_test Validation_47334256_112063.700_00004119
720
+ ARKitScenes_test Validation_47334256_112064.699_00004179
721
+ ARKitScenes_test Validation_47334256_112065.699_00004239
722
+ ARKitScenes_test Validation_47429922_136624.256_00000517
723
+ ARKitScenes_test Validation_47429922_136657.242_00002497
724
+ ARKitScenes_test Validation_47429971_16349.760_00004238
725
+ ARKitScenes_test Validation_47430002_17171.758_00000277
726
+ ARKitScenes_test Validation_47430005_17517.360_00000276
727
+ ARKitScenes_test Validation_47430005_17532.371_00001177
728
+ ARKitScenes_test Validation_47430023_18906.937_00002864
729
+ ARKitScenes_test Validation_47430033_19414.387_00000403
730
+ ARKitScenes_test Validation_47430038_20536.784_00000216
731
+ ARKitScenes_test Validation_47430043_20606.288_00001717
732
+ ARKitScenes_test Validation_47430047_21122.842_00001898
733
+ ARKitScenes_test Validation_47895341_96403.324_00001057
734
+ ARKitScenes_test Validation_47895341_96411.321_00001537
735
+ ARKitScenes_test Validation_47895348_96488.523_00001777
736
+ ARKitScenes_test Validation_47895348_96538.520_00004778
737
+ ARKitScenes_test Validation_47895350_96557.812_00000396
738
+ ARKitScenes_test Validation_47895350_96559.828_00000517
739
+ ARKitScenes_test Validation_47895350_96571.823_00001237
740
+ ARKitScenes_test Validation_47895350_96604.826_00003218
741
+ ARKitScenes_test Validation_47895350_96610.824_00003578
742
+ ARKitScenes_test Validation_47895350_96621.819_00004238
743
+ ARKitScenes_test Validation_47895350_96646.826_00005739
744
+ ARKitScenes_test Validation_47895350_96659.821_00006519
745
+ ARKitScenes_test Validation_47895350_96663.819_00006759
746
+ ARKitScenes_test Validation_47895350_96730.825_00010781
747
+ ARKitScenes_test Validation_47895355_98775.628_00001297
748
+ ARKitScenes_test Validation_47895355_98788.639_00002078
749
+ ARKitScenes_test Validation_47895355_98796.636_00002558
750
+ ARKitScenes_test Validation_47895355_98810.630_00003398
751
+ ARKitScenes_test Validation_47895355_98823.642_00004179
752
+ ARKitScenes_test Validation_47895355_98827.640_00004419
753
+ ARKitScenes_test Validation_47895355_98906.624_00009160
754
+ ARKitScenes_test Validation_47895355_98910.639_00009401
755
+ ARKitScenes_test Validation_47895355_98929.632_00010541
756
+ ARKitScenes_test Validation_47895355_98943.626_00011381
757
+ ARKitScenes_test Validation_47895355_98966.633_00012762
758
+ ARKitScenes_test Validation_47895364_100221.941_00003158
759
+ ARKitScenes_test Validation_47895364_100251.946_00004959
760
+ ARKitScenes_test Validation_47895364_100287.931_00007119
761
+ ARKitScenes_test Validation_47895364_100307.940_00008320
762
+ ARKitScenes_test Validation_48018367_389452.575_00000457
763
+ ARKitScenes_test Validation_48018367_389473.583_00001718
764
+ ARKitScenes_test Validation_48018368_389553.283_00001190
765
+ ARKitScenes_test Validation_48018368_389572.275_00002330
766
+ ARKitScenes_test Validation_48018368_389581.271_00002870
767
+ ARKitScenes_test Validation_48018372_389531.675_00002498
768
+ ARKitScenes_test Validation_48018379_390040.136_00002678
769
+ ARKitScenes_test Validation_48018559_3346.848_00000337
770
+ ARKitScenes_test Validation_48018566_3858.438_00000103
771
+ ARKitScenes_test Validation_48018571_3905.036_00000439
772
+ ARKitScenes_test Validation_48018572_3798.046_00000128
773
+ ARKitScenes_test Validation_48018572_3810.041_00000848
774
+ ARKitScenes_test Validation_48018730_1678.470_00000637
775
+ ARKitScenes_test Validation_48018730_1723.469_00003338
776
+ ARKitScenes_test Validation_48018730_1769.467_00006099
777
+ ARKitScenes_test Validation_48018732_1939.564_00001759
778
+ ARKitScenes_test Validation_48018732_1942.563_00001939
779
+ ARKitScenes_test Validation_48018732_1945.579_00002120
780
+ ARKitScenes_test Validation_48018732_1948.577_00002300
781
+ ARKitScenes_test Validation_48018732_1953.575_00002600
782
+ ARKitScenes_test Validation_48018741_3393.825_00000337
783
+ ARKitScenes_test Validation_48458481_111522.301_00001117
784
+ ARKitScenes_test Validation_48458484_111590.308_00001237
785
+ ARKitScenes_test Validation_48458489_111549.307_00000703
786
+ ARKitScenes_test Validation_48458489_111557.304_00001183
787
+ ARKitScenes_test Validation_48458650_7066.602_00000456
788
+ ARKitScenes_test Validation_48458650_7068.601_00000576
789
+ ARKitScenes_test Validation_48458650_7070.617_00000697
790
+ ARKitScenes_test Validation_48458650_7071.617_00000757
791
+ ARKitScenes_test Validation_48458656_7661.589_00002077
792
+ ARKitScenes_test Validation_48458656_7663.589_00002197
793
+ ARKitScenes_test Validation_48458656_7667.587_00002437
794
+ ARKitScenes_test Validation_48458656_7672.585_00002737
795
+ ARKitScenes_test Validation_48458657_7617.691_00002023
796
+ ARKitScenes_test Validation_48458660_8270.961_00001658
797
+ ARKitScenes_test Validation_48458667_8365.655_00000637
798
+ ARKitScenes_test Validation_48458667_8367.654_00000757
799
+ ARKitScenes_test Validation_48458667_8376.651_00001297
800
+ ARKitScenes_test Validation_48458667_8395.660_00002438
external/WildCamera/splits/biwirgbdid_test.txt ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ BIWIRGBDID Testing_Still_000_000a_000018-a_111628807_rgb
2
+ BIWIRGBDID Testing_Still_000_000a_000019-a_111628891_rgb
3
+ BIWIRGBDID Testing_Still_000_000a_000124-a_111639723_rgb
4
+ BIWIRGBDID Testing_Still_000_000a_000157-a_111643139_rgb
5
+ BIWIRGBDID Testing_Still_000_000a_000179-a_111645307_rgb
6
+ BIWIRGBDID Testing_Still_000_000a_000183-a_111645807_rgb
7
+ BIWIRGBDID Testing_Still_000_000a_000195-a_111647059_rgb
8
+ BIWIRGBDID Testing_Still_000_000a_000200-a_111647559_rgb
9
+ BIWIRGBDID Testing_Still_001_001a_000022-a_106257171_rgb
10
+ BIWIRGBDID Testing_Still_001_001a_000039-a_106258923_rgb
11
+ BIWIRGBDID Testing_Still_001_001a_000056-a_106260671_rgb
12
+ BIWIRGBDID Testing_Still_001_001a_000090-a_106264423_rgb
13
+ BIWIRGBDID Testing_Still_001_001a_000099-a_106265423_rgb
14
+ BIWIRGBDID Testing_Still_001_001a_000117-a_106267255_rgb
15
+ BIWIRGBDID Testing_Still_001_001a_000145-a_106270255_rgb
16
+ BIWIRGBDID Testing_Still_001_001a_000157-a_106271503_rgb
17
+ BIWIRGBDID Testing_Still_002_002a_000016-a_1232924843_rgb
18
+ BIWIRGBDID Testing_Still_002_002a_000025-a_1232925843_rgb
19
+ BIWIRGBDID Testing_Still_002_002a_000035-a_1232926927_rgb
20
+ BIWIRGBDID Testing_Still_002_002a_000069-a_1232930595_rgb
21
+ BIWIRGBDID Testing_Still_002_002a_000088-a_1232932679_rgb
22
+ BIWIRGBDID Testing_Still_002_002a_000089-a_1232932759_rgb
23
+ BIWIRGBDID Testing_Still_002_002a_000092-a_1232933095_rgb
24
+ BIWIRGBDID Testing_Still_002_002a_000094-a_1232933259_rgb
25
+ BIWIRGBDID Testing_Still_002_002a_000123-a_1232936595_rgb
26
+ BIWIRGBDID Testing_Still_002_002a_000130-a_1232937342_rgb
27
+ BIWIRGBDID Testing_Still_002_002a_000141-a_1232938510_rgb
28
+ BIWIRGBDID Testing_Still_002_002a_000143-a_1232938759_rgb
29
+ BIWIRGBDID Testing_Still_002_002a_000146-a_1232939010_rgb
30
+ BIWIRGBDID Testing_Still_003_003a_000006-a_621308413_rgb
31
+ BIWIRGBDID Testing_Still_003_003a_000007-a_621308581_rgb
32
+ BIWIRGBDID Testing_Still_003_003a_000033-a_621311497_rgb
33
+ BIWIRGBDID Testing_Still_003_003a_000050-a_621313413_rgb
34
+ BIWIRGBDID Testing_Still_003_003a_000110-a_621320413_rgb
35
+ BIWIRGBDID Testing_Still_003_003a_000141-a_621324749_rgb
36
+ BIWIRGBDID Testing_Still_006_006a_000000-a_1224024579_rgb
37
+ BIWIRGBDID Testing_Still_006_006a_000002-a_1224024747_rgb
38
+ BIWIRGBDID Testing_Still_006_006a_000034-a_1224028080_rgb
39
+ BIWIRGBDID Testing_Still_006_006a_000055-a_1224030247_rgb
40
+ BIWIRGBDID Testing_Still_006_006a_000070-a_1224031831_rgb
41
+ BIWIRGBDID Testing_Still_006_006a_000089-a_1224033747_rgb
42
+ BIWIRGBDID Testing_Still_006_006a_000092-a_1224034079_rgb
43
+ BIWIRGBDID Testing_Still_006_006a_000098-a_1224034663_rgb
44
+ BIWIRGBDID Testing_Still_006_006a_000106-a_1224035495_rgb
45
+ BIWIRGBDID Testing_Still_006_006a_000122-a_1224037163_rgb
46
+ BIWIRGBDID Testing_Still_006_006a_000124-a_1224037331_rgb
47
+ BIWIRGBDID Testing_Still_006_006a_000139-a_1224038911_rgb
48
+ BIWIRGBDID Testing_Still_006_006a_000151-a_1224040163_rgb
49
+ BIWIRGBDID Testing_Still_008_008a_000014-a_190303117_rgb
50
+ BIWIRGBDID Testing_Still_008_008a_000029-a_190304618_rgb
51
+ BIWIRGBDID Testing_Still_008_008a_000073-a_190309201_rgb
52
+ BIWIRGBDID Testing_Still_008_008a_000092-a_190311285_rgb
53
+ BIWIRGBDID Testing_Still_008_008a_000116-a_190313701_rgb
54
+ BIWIRGBDID Testing_Still_008_008a_000125-a_190314617_rgb
55
+ BIWIRGBDID Testing_Still_010_010a_000005-a_171989565_rgb
56
+ BIWIRGBDID Testing_Still_010_010a_000019-a_171990981_rgb
57
+ BIWIRGBDID Testing_Still_010_010a_000021-a_171991229_rgb
58
+ BIWIRGBDID Testing_Still_010_010a_000054-a_171994564_rgb
59
+ BIWIRGBDID Testing_Still_010_010a_000063-a_171995649_rgb
60
+ BIWIRGBDID Testing_Still_010_010a_000072-a_171996480_rgb
61
+ BIWIRGBDID Testing_Still_010_010a_000085-a_171997897_rgb
62
+ BIWIRGBDID Testing_Still_010_010a_000087-a_171998064_rgb
63
+ BIWIRGBDID Testing_Still_010_010a_000098-a_171999229_rgb
64
+ BIWIRGBDID Testing_Still_010_010a_000107-a_172000148_rgb
65
+ BIWIRGBDID Testing_Still_010_010a_000119-a_172001396_rgb
66
+ BIWIRGBDID Testing_Still_011_011a_000023-a_200194778_rgb
67
+ BIWIRGBDID Testing_Still_011_011a_000088-a_200201941_rgb
68
+ BIWIRGBDID Testing_Still_011_011a_000101-a_200203442_rgb
69
+ BIWIRGBDID Testing_Still_011_011a_000104-a_200203778_rgb
70
+ BIWIRGBDID Testing_Still_011_011a_000116-a_200205277_rgb
71
+ BIWIRGBDID Testing_Still_016_016a_000000-a_188165678_rgb
72
+ BIWIRGBDID Testing_Still_016_016a_000026-a_188168426_rgb
73
+ BIWIRGBDID Testing_Still_016_016a_000043-a_188170258_rgb
74
+ BIWIRGBDID Testing_Still_016_016a_000052-a_188171178_rgb
75
+ BIWIRGBDID Testing_Still_016_016a_000114-a_188177758_rgb
76
+ BIWIRGBDID Testing_Still_017_017a_000004-a_191551441_rgb
77
+ BIWIRGBDID Testing_Still_017_017a_000050-a_191556773_rgb
78
+ BIWIRGBDID Testing_Still_017_017a_000055-a_191557272_rgb
79
+ BIWIRGBDID Testing_Still_017_017a_000059-a_191557688_rgb
80
+ BIWIRGBDID Testing_Still_017_017a_000073-a_191559189_rgb
81
+ BIWIRGBDID Testing_Still_017_017a_000099-a_191562440_rgb
82
+ BIWIRGBDID Testing_Still_017_017a_000102-a_191562688_rgb
83
+ BIWIRGBDID Testing_Still_017_017a_000103-a_191562772_rgb
84
+ BIWIRGBDID Testing_Still_017_017a_000122-a_191564772_rgb
85
+ BIWIRGBDID Testing_Still_019_019a_000024-a_107178865_rgb
86
+ BIWIRGBDID Testing_Still_019_019a_000081-a_107184780_rgb
87
+ BIWIRGBDID Testing_Still_019_019a_000086-a_107185364_rgb
88
+ BIWIRGBDID Testing_Still_021_021a_000028-a_109006730_rgb
89
+ BIWIRGBDID Testing_Still_021_021a_000081-a_109012398_rgb
90
+ BIWIRGBDID Testing_Still_021_021a_000084-a_109012646_rgb
91
+ BIWIRGBDID Testing_Still_021_021a_000100-a_109014314_rgb
92
+ BIWIRGBDID Testing_Still_021_021a_000117-a_109016062_rgb
93
+ BIWIRGBDID Testing_Still_021_021a_000137-a_109018146_rgb
94
+ BIWIRGBDID Testing_Still_022_022a_000052-a_1214872830_rgb
95
+ BIWIRGBDID Testing_Still_022_022a_000053-a_1214872914_rgb
96
+ BIWIRGBDID Testing_Still_022_022a_000072-a_1214875162_rgb
97
+ BIWIRGBDID Testing_Still_023_023a_000003-a_197683355_rgb
98
+ BIWIRGBDID Testing_Still_023_023a_000012-a_197684603_rgb
99
+ BIWIRGBDID Testing_Still_023_023a_000024-a_197686103_rgb
100
+ BIWIRGBDID Testing_Still_023_023a_000027-a_197686523_rgb
101
+ BIWIRGBDID Testing_Still_023_023a_000096-a_197695603_rgb
102
+ BIWIRGBDID Testing_Still_023_023a_000128-a_197699687_rgb
103
+ BIWIRGBDID Testing_Still_023_023a_000138-a_197700939_rgb
104
+ BIWIRGBDID Testing_Still_024_024a_000027-a_197310296_rgb
105
+ BIWIRGBDID Testing_Still_024_024a_000033-a_197311045_rgb
106
+ BIWIRGBDID Testing_Still_024_024a_000046-a_197312628_rgb
107
+ BIWIRGBDID Testing_Still_024_024a_000054-a_197313712_rgb
108
+ BIWIRGBDID Testing_Still_024_024a_000076-a_197316628_rgb
109
+ BIWIRGBDID Testing_Still_024_024a_000093-a_197319128_rgb
110
+ BIWIRGBDID Testing_Still_026_026a_000001-a_110999274_rgb
111
+ BIWIRGBDID Testing_Still_026_026a_000003-a_110999522_rgb
112
+ BIWIRGBDID Testing_Still_026_026a_000017-a_111001106_rgb
113
+ BIWIRGBDID Testing_Still_026_026a_000027-a_111002106_rgb
114
+ BIWIRGBDID Testing_Still_026_026a_000029-a_111002354_rgb
115
+ BIWIRGBDID Testing_Still_026_026a_000072-a_111007106_rgb
116
+ BIWIRGBDID Testing_Still_026_026a_000077-a_111007606_rgb
117
+ BIWIRGBDID Testing_Still_026_026a_000110-a_111011022_rgb
118
+ BIWIRGBDID Testing_Still_026_026a_000112-a_111011270_rgb
119
+ BIWIRGBDID Testing_Still_027_027a_000012-a_111337418_rgb
120
+ BIWIRGBDID Testing_Still_027_027a_000018-a_111338086_rgb
121
+ BIWIRGBDID Testing_Still_027_027a_000031-a_111339750_rgb
122
+ BIWIRGBDID Testing_Still_027_027a_000033-a_111339918_rgb
123
+ BIWIRGBDID Testing_Still_027_027a_000044-a_111341169_rgb
124
+ BIWIRGBDID Testing_Still_027_027a_000045-a_111341333_rgb
125
+ BIWIRGBDID Testing_Still_027_027a_000119-a_111349249_rgb
126
+ BIWIRGBDID Testing_Still_027_027a_000134-a_111350917_rgb
127
+ BIWIRGBDID Testing_Still_027_027a_000142-a_111351749_rgb
128
+ BIWIRGBDID Testing_Still_030_030a_000012-a_1233788639_rgb
129
+ BIWIRGBDID Testing_Still_030_030a_000043-a_1233791891_rgb
130
+ BIWIRGBDID Testing_Still_030_030a_000072-a_1233795471_rgb
131
+ BIWIRGBDID Testing_Still_030_030a_000091-a_1233797723_rgb
132
+ BIWIRGBDID Testing_Still_030_030a_000096-a_1233798307_rgb
133
+ BIWIRGBDID Testing_Still_030_030a_000107-a_1233799471_rgb
134
+ BIWIRGBDID Testing_Still_030_030a_000120-a_1233800971_rgb
135
+ BIWIRGBDID Testing_Still_030_030a_000127-a_1233801807_rgb
136
+ BIWIRGBDID Testing_Still_031_031a_000004-a_186779771_rgb
137
+ BIWIRGBDID Testing_Still_031_031a_000014-a_186780855_rgb
138
+ BIWIRGBDID Testing_Still_031_031a_000036-a_186783439_rgb
139
+ BIWIRGBDID Testing_Still_031_031a_000050-a_186784855_rgb
140
+ BIWIRGBDID Testing_Still_031_031a_000086-a_186788523_rgb
141
+ BIWIRGBDID Testing_Still_031_031a_000098-a_186789771_rgb
142
+ BIWIRGBDID Testing_Still_031_031a_000102-a_186790186_rgb
143
+ BIWIRGBDID Testing_Still_031_031a_000120-a_186791938_rgb
144
+ BIWIRGBDID Testing_Still_032_032a_000032-a_111072184_rgb
145
+ BIWIRGBDID Testing_Still_032_032a_000049-a_111074016_rgb
146
+ BIWIRGBDID Testing_Still_032_032a_000112-a_111080852_rgb
147
+ BIWIRGBDID Testing_Still_032_032a_000122-a_111081932_rgb
148
+ BIWIRGBDID Testing_Still_036_036a_000011-a_107768409_rgb
149
+ BIWIRGBDID Testing_Still_036_036a_000016-a_107768826_rgb
150
+ BIWIRGBDID Testing_Still_036_036a_000024-a_107769661_rgb
151
+ BIWIRGBDID Testing_Still_036_036a_000071-a_107774662_rgb
152
+ BIWIRGBDID Testing_Still_036_036a_000077-a_107775325_rgb
153
+ BIWIRGBDID Testing_Still_036_036a_000087-a_107776409_rgb
154
+ BIWIRGBDID Testing_Still_037_037a_000040-a_191434279_rgb
155
+ BIWIRGBDID Testing_Still_037_037a_000044-a_191434696_rgb
156
+ BIWIRGBDID Testing_Still_037_037a_000074-a_191438032_rgb
157
+ BIWIRGBDID Testing_Still_037_037a_000091-a_191439695_rgb
158
+ BIWIRGBDID Testing_Still_037_037a_000109-a_191441696_rgb
159
+ BIWIRGBDID Testing_Still_037_037a_000111-a_191442031_rgb
160
+ BIWIRGBDID Testing_Still_037_037a_000123-a_191443363_rgb
161
+ BIWIRGBDID Testing_Still_039_039a_000006-a_1219302217_rgb
162
+ BIWIRGBDID Testing_Still_039_039a_000015-a_1219303217_rgb
163
+ BIWIRGBDID Testing_Still_039_039a_000044-a_1219306217_rgb
164
+ BIWIRGBDID Testing_Still_039_039a_000058-a_1219307717_rgb
165
+ BIWIRGBDID Testing_Still_039_039a_000090-a_1219310969_rgb
166
+ BIWIRGBDID Testing_Still_039_039a_000097-a_1219311885_rgb
167
+ BIWIRGBDID Testing_Still_039_039a_000122-a_1219314717_rgb
168
+ BIWIRGBDID Testing_Still_039_039a_000144-a_1219317049_rgb
169
+ BIWIRGBDID Testing_Still_040_040a_000040-a_1231702678_rgb
170
+ BIWIRGBDID Testing_Still_040_040a_000046-a_1231703346_rgb
171
+ BIWIRGBDID Testing_Still_040_040a_000073-a_1231706514_rgb
172
+ BIWIRGBDID Testing_Still_040_040a_000102-a_1231710014_rgb
173
+ BIWIRGBDID Testing_Still_040_040a_000121-a_1231712093_rgb
174
+ BIWIRGBDID Testing_Still_040_040a_000122-a_1231712177_rgb
175
+ BIWIRGBDID Testing_Still_040_040a_000136-a_1231713846_rgb
176
+ BIWIRGBDID Testing_Still_044_044a_000002-a_1214165465_rgb
177
+ BIWIRGBDID Testing_Still_044_044a_000046-a_1214170130_rgb
178
+ BIWIRGBDID Testing_Still_044_044a_000059-a_1214171381_rgb
179
+ BIWIRGBDID Testing_Still_044_044a_000070-a_1214172549_rgb
180
+ BIWIRGBDID Testing_Still_044_044a_000089-a_1214174465_rgb
181
+ BIWIRGBDID Testing_Still_044_044a_000103-a_1214175881_rgb
182
+ BIWIRGBDID Testing_Still_046_046a_000006-a_111209841_rgb
183
+ BIWIRGBDID Testing_Still_046_046a_000021-a_111211425_rgb
184
+ BIWIRGBDID Testing_Still_046_046a_000040-a_111213425_rgb
185
+ BIWIRGBDID Testing_Still_046_046a_000071-a_111216841_rgb
186
+ BIWIRGBDID Testing_Still_046_046a_000078-a_111217509_rgb
187
+ BIWIRGBDID Testing_Still_046_046a_000079-a_111217676_rgb
188
+ BIWIRGBDID Testing_Still_047_047a_000067-a_1214094887_rgb
189
+ BIWIRGBDID Testing_Still_047_047a_000077-a_1214095887_rgb
190
+ BIWIRGBDID Testing_Still_047_047a_000090-a_1214097135_rgb
191
+ BIWIRGBDID Testing_Still_047_047a_000120-a_1214100468_rgb
192
+ BIWIRGBDID Testing_Still_049_049a_000000-a_201357193_rgb
193
+ BIWIRGBDID Testing_Still_049_049a_000020-a_201359362_rgb
194
+ BIWIRGBDID Testing_Still_049_049a_000026-a_201360029_rgb
195
+ BIWIRGBDID Testing_Still_049_049a_000028-a_201360193_rgb
196
+ BIWIRGBDID Testing_Still_049_049a_000069-a_201364613_rgb
197
+ BIWIRGBDID Testing_Still_049_049a_000073-a_201365029_rgb
198
+ BIWIRGBDID Testing_Still_049_049a_000083-a_201366113_rgb
199
+ BIWIRGBDID Testing_Still_049_049a_000088-a_201366613_rgb
200
+ BIWIRGBDID Testing_Walking_000_000b_000008-a_111676306_rgb
201
+ BIWIRGBDID Testing_Walking_000_000b_000032-a_111678306_rgb
202
+ BIWIRGBDID Testing_Walking_000_000b_000043-a_111679306_rgb
203
+ BIWIRGBDID Testing_Walking_000_000b_000064-a_111681390_rgb
204
+ BIWIRGBDID Testing_Walking_000_000b_000066-a_111681638_rgb
205
+ BIWIRGBDID Testing_Walking_000_000b_000085-a_111683806_rgb
206
+ BIWIRGBDID Testing_Walking_000_000b_000110-a_111686054_rgb
207
+ BIWIRGBDID Testing_Walking_000_000b_000139-a_111688470_rgb
208
+ BIWIRGBDID Testing_Walking_000_000b_000158-a_111690054_rgb
209
+ BIWIRGBDID Testing_Walking_000_000b_000159-a_111690138_rgb
210
+ BIWIRGBDID Testing_Walking_000_000b_000182-a_111692390_rgb
211
+ BIWIRGBDID Testing_Walking_000_000b_000192-a_111693390_rgb
212
+ BIWIRGBDID Testing_Walking_000_000b_000234-a_111697222_rgb
213
+ BIWIRGBDID Testing_Walking_000_000b_000258-a_111699221_rgb
214
+ BIWIRGBDID Testing_Walking_000_000b_000261-a_111699469_rgb
215
+ BIWIRGBDID Testing_Walking_000_000b_000265-a_111699886_rgb
216
+ BIWIRGBDID Testing_Walking_000_000b_000290-a_111702469_rgb
217
+ BIWIRGBDID Testing_Walking_000_000b_000291-a_111702553_rgb
218
+ BIWIRGBDID Testing_Walking_000_000b_000349-a_111708305_rgb
219
+ BIWIRGBDID Testing_Walking_000_000b_000355-a_111708805_rgb
220
+ BIWIRGBDID Testing_Walking_000_000b_000370-a_111710137_rgb
221
+ BIWIRGBDID Testing_Walking_000_000b_000372-a_111710305_rgb
222
+ BIWIRGBDID Testing_Walking_000_000b_000406-a_111713721_rgb
223
+ BIWIRGBDID Testing_Walking_000_000b_000410-a_111714137_rgb
224
+ BIWIRGBDID Testing_Walking_000_000b_000416-a_111714637_rgb
225
+ BIWIRGBDID Testing_Walking_000_000b_000420-a_111714969_rgb
226
+ BIWIRGBDID Testing_Walking_000_000b_000439-a_111716637_rgb
227
+ BIWIRGBDID Testing_Walking_001_001b_000002-a_106302670_rgb
228
+ BIWIRGBDID Testing_Walking_001_001b_000004-a_106302835_rgb
229
+ BIWIRGBDID Testing_Walking_001_001b_000011-a_106303418_rgb
230
+ BIWIRGBDID Testing_Walking_001_001b_000015-a_106303834_rgb
231
+ BIWIRGBDID Testing_Walking_001_001b_000030-a_106305254_rgb
232
+ BIWIRGBDID Testing_Walking_001_001b_000046-a_106306754_rgb
233
+ BIWIRGBDID Testing_Walking_001_001b_000083-a_106310586_rgb
234
+ BIWIRGBDID Testing_Walking_001_001b_000158-a_106317918_rgb
235
+ BIWIRGBDID Testing_Walking_001_001b_000177-a_106319586_rgb
236
+ BIWIRGBDID Testing_Walking_001_001b_000181-a_106319918_rgb
237
+ BIWIRGBDID Testing_Walking_001_001b_000205-a_106322170_rgb
238
+ BIWIRGBDID Testing_Walking_001_001b_000249-a_106327250_rgb
239
+ BIWIRGBDID Testing_Walking_001_001b_000289-a_106331834_rgb
240
+ BIWIRGBDID Testing_Walking_001_001b_000303-a_106333334_rgb
241
+ BIWIRGBDID Testing_Walking_001_001b_000329-a_106336502_rgb
242
+ BIWIRGBDID Testing_Walking_001_001b_000357-a_106339586_rgb
243
+ BIWIRGBDID Testing_Walking_001_001b_000360-a_106339917_rgb
244
+ BIWIRGBDID Testing_Walking_001_001b_000371-a_106341165_rgb
245
+ BIWIRGBDID Testing_Walking_002_002b_000004-a_1233001421_rgb
246
+ BIWIRGBDID Testing_Walking_002_002b_000018-a_1233002673_rgb
247
+ BIWIRGBDID Testing_Walking_002_002b_000024-a_1233003257_rgb
248
+ BIWIRGBDID Testing_Walking_002_002b_000032-a_1233004005_rgb
249
+ BIWIRGBDID Testing_Walking_002_002b_000059-a_1233006589_rgb
250
+ BIWIRGBDID Testing_Walking_002_002b_000069-a_1233007673_rgb
251
+ BIWIRGBDID Testing_Walking_002_002b_000130-a_1233014089_rgb
252
+ BIWIRGBDID Testing_Walking_002_002b_000159-a_1233017257_rgb
253
+ BIWIRGBDID Testing_Walking_002_002b_000169-a_1233018337_rgb
254
+ BIWIRGBDID Testing_Walking_002_002b_000170-a_1233018421_rgb
255
+ BIWIRGBDID Testing_Walking_002_002b_000188-a_1233020337_rgb
256
+ BIWIRGBDID Testing_Walking_002_002b_000220-a_1233024004_rgb
257
+ BIWIRGBDID Testing_Walking_002_002b_000222-a_1233024172_rgb
258
+ BIWIRGBDID Testing_Walking_002_002b_000232-a_1233025336_rgb
259
+ BIWIRGBDID Testing_Walking_002_002b_000287-a_1233031336_rgb
260
+ BIWIRGBDID Testing_Walking_002_002b_000303-a_1233033088_rgb
261
+ BIWIRGBDID Testing_Walking_002_002b_000308-a_1233033504_rgb
262
+ BIWIRGBDID Testing_Walking_002_002b_000318-a_1233034420_rgb
263
+ BIWIRGBDID Testing_Walking_002_002b_000338-a_1233036588_rgb
264
+ BIWIRGBDID Testing_Walking_002_002b_000340-a_1233036836_rgb
265
+ BIWIRGBDID Testing_Walking_002_002b_000355-a_1233038504_rgb
266
+ BIWIRGBDID Testing_Walking_003_003b_000025-a_621819949_rgb
267
+ BIWIRGBDID Testing_Walking_003_003b_000054-a_621823117_rgb
268
+ BIWIRGBDID Testing_Walking_003_003b_000112-a_621829533_rgb
269
+ BIWIRGBDID Testing_Walking_003_003b_000113-a_621829617_rgb
270
+ BIWIRGBDID Testing_Walking_003_003b_000173-a_621836365_rgb
271
+ BIWIRGBDID Testing_Walking_003_003b_000179-a_621836949_rgb
272
+ BIWIRGBDID Testing_Walking_003_003b_000189-a_621837781_rgb
273
+ BIWIRGBDID Testing_Walking_003_003b_000191-a_621837949_rgb
274
+ BIWIRGBDID Testing_Walking_003_003b_000225-a_621842201_rgb
275
+ BIWIRGBDID Testing_Walking_003_003b_000274-a_621847616_rgb
276
+ BIWIRGBDID Testing_Walking_003_003b_000275-a_621847701_rgb
277
+ BIWIRGBDID Testing_Walking_003_003b_000278-a_621847948_rgb
278
+ BIWIRGBDID Testing_Walking_003_003b_000306-a_621851116_rgb
279
+ BIWIRGBDID Testing_Walking_003_003b_000323-a_621852948_rgb
280
+ BIWIRGBDID Testing_Walking_003_003b_000335-a_621854032_rgb
281
+ BIWIRGBDID Testing_Walking_006_006b_000047-a_1224090658_rgb
282
+ BIWIRGBDID Testing_Walking_006_006b_000048-a_1224090742_rgb
283
+ BIWIRGBDID Testing_Walking_006_006b_000053-a_1224091325_rgb
284
+ BIWIRGBDID Testing_Walking_006_006b_000080-a_1224093993_rgb
285
+ BIWIRGBDID Testing_Walking_006_006b_000106-a_1224096741_rgb
286
+ BIWIRGBDID Testing_Walking_006_006b_000112-a_1224097325_rgb
287
+ BIWIRGBDID Testing_Walking_006_006b_000118-a_1224097825_rgb
288
+ BIWIRGBDID Testing_Walking_006_006b_000119-a_1224097909_rgb
289
+ BIWIRGBDID Testing_Walking_006_006b_000127-a_1224098657_rgb
290
+ BIWIRGBDID Testing_Walking_006_006b_000158-a_1224101325_rgb
291
+ BIWIRGBDID Testing_Walking_006_006b_000168-a_1224102241_rgb
292
+ BIWIRGBDID Testing_Walking_006_006b_000205-a_1224106073_rgb
293
+ BIWIRGBDID Testing_Walking_006_006b_000236-a_1224108825_rgb
294
+ BIWIRGBDID Testing_Walking_006_006b_000257-a_1224110573_rgb
295
+ BIWIRGBDID Testing_Walking_006_006b_000264-a_1224111157_rgb
296
+ BIWIRGBDID Testing_Walking_006_006b_000289-a_1224113825_rgb
297
+ BIWIRGBDID Testing_Walking_006_006b_000396-a_1224124073_rgb
298
+ BIWIRGBDID Testing_Walking_008_008b_000001-a_190329700_rgb
299
+ BIWIRGBDID Testing_Walking_008_008b_000030-a_190332200_rgb
300
+ BIWIRGBDID Testing_Walking_008_008b_000065-a_190336032_rgb
301
+ BIWIRGBDID Testing_Walking_008_008b_000087-a_190338532_rgb
302
+ BIWIRGBDID Testing_Walking_008_008b_000090-a_190338865_rgb
303
+ BIWIRGBDID Testing_Walking_008_008b_000093-a_190339117_rgb
304
+ BIWIRGBDID Testing_Walking_008_008b_000103-a_190339948_rgb
305
+ BIWIRGBDID Testing_Walking_008_008b_000124-a_190341700_rgb
306
+ BIWIRGBDID Testing_Walking_008_008b_000170-a_190346116_rgb
307
+ BIWIRGBDID Testing_Walking_008_008b_000218-a_190351280_rgb
308
+ BIWIRGBDID Testing_Walking_008_008b_000231-a_190352364_rgb
309
+ BIWIRGBDID Testing_Walking_008_008b_000257-a_190354780_rgb
310
+ BIWIRGBDID Testing_Walking_008_008b_000268-a_190355700_rgb
311
+ BIWIRGBDID Testing_Walking_008_008b_000307-a_190359696_rgb
312
+ BIWIRGBDID Testing_Walking_008_008b_000309-a_190359864_rgb
313
+ BIWIRGBDID Testing_Walking_008_008b_000317-a_190360531_rgb
314
+ BIWIRGBDID Testing_Walking_008_008b_000329-a_190361615_rgb
315
+ BIWIRGBDID Testing_Walking_008_008b_000363-a_190364948_rgb
316
+ BIWIRGBDID Testing_Walking_010_010b_000011-a_172014564_rgb
317
+ BIWIRGBDID Testing_Walking_010_010b_000027-a_172015896_rgb
318
+ BIWIRGBDID Testing_Walking_010_010b_000030-a_172016144_rgb
319
+ BIWIRGBDID Testing_Walking_010_010b_000077-a_172020812_rgb
320
+ BIWIRGBDID Testing_Walking_010_010b_000142-a_172026812_rgb
321
+ BIWIRGBDID Testing_Walking_010_010b_000167-a_172029395_rgb
322
+ BIWIRGBDID Testing_Walking_010_010b_000176-a_172030396_rgb
323
+ BIWIRGBDID Testing_Walking_010_010b_000187-a_172031560_rgb
324
+ BIWIRGBDID Testing_Walking_010_010b_000212-a_172033980_rgb
325
+ BIWIRGBDID Testing_Walking_010_010b_000237-a_172036060_rgb
326
+ BIWIRGBDID Testing_Walking_010_010b_000255-a_172037727_rgb
327
+ BIWIRGBDID Testing_Walking_010_010b_000282-a_172040395_rgb
328
+ BIWIRGBDID Testing_Walking_010_010b_000321-a_172044395_rgb
329
+ BIWIRGBDID Testing_Walking_010_010b_000324-a_172044643_rgb
330
+ BIWIRGBDID Testing_Walking_010_010b_000330-a_172045311_rgb
331
+ BIWIRGBDID Testing_Walking_010_010b_000351-a_172047143_rgb
332
+ BIWIRGBDID Testing_Walking_010_010b_000352-a_172047227_rgb
333
+ BIWIRGBDID Testing_Walking_010_010b_000380-a_172049643_rgb
334
+ BIWIRGBDID Testing_Walking_010_010b_000382-a_172049811_rgb
335
+ BIWIRGBDID Testing_Walking_011_011b_000025-a_200221193_rgb
336
+ BIWIRGBDID Testing_Walking_011_011b_000033-a_200222025_rgb
337
+ BIWIRGBDID Testing_Walking_011_011b_000044-a_200223193_rgb
338
+ BIWIRGBDID Testing_Walking_011_011b_000049-a_200223693_rgb
339
+ BIWIRGBDID Testing_Walking_011_011b_000083-a_200227525_rgb
340
+ BIWIRGBDID Testing_Walking_011_011b_000086-a_200227857_rgb
341
+ BIWIRGBDID Testing_Walking_011_011b_000106-a_200230193_rgb
342
+ BIWIRGBDID Testing_Walking_011_011b_000125-a_200232025_rgb
343
+ BIWIRGBDID Testing_Walking_011_011b_000142-a_200233609_rgb
344
+ BIWIRGBDID Testing_Walking_011_011b_000175-a_200237109_rgb
345
+ BIWIRGBDID Testing_Walking_011_011b_000203-a_200240357_rgb
346
+ BIWIRGBDID Testing_Walking_011_011b_000204-a_200240441_rgb
347
+ BIWIRGBDID Testing_Walking_011_011b_000259-a_200246188_rgb
348
+ BIWIRGBDID Testing_Walking_011_011b_000266-a_200246857_rgb
349
+ BIWIRGBDID Testing_Walking_011_011b_000293-a_200249524_rgb
350
+ BIWIRGBDID Testing_Walking_011_011b_000301-a_200250440_rgb
351
+ BIWIRGBDID Testing_Walking_011_011b_000309-a_200251441_rgb
352
+ BIWIRGBDID Testing_Walking_011_011b_000321-a_200252772_rgb
353
+ BIWIRGBDID Testing_Walking_011_011b_000367-a_200257440_rgb
354
+ BIWIRGBDID Testing_Walking_011_011b_000393-a_200259772_rgb
355
+ BIWIRGBDID Testing_Walking_011_011b_000398-a_200260272_rgb
356
+ BIWIRGBDID Testing_Walking_011_011b_000430-a_200263772_rgb
357
+ BIWIRGBDID Testing_Walking_016_016b_000003-a_188187593_rgb
358
+ BIWIRGBDID Testing_Walking_016_016b_000004-a_188187673_rgb
359
+ BIWIRGBDID Testing_Walking_016_016b_000016-a_188188841_rgb
360
+ BIWIRGBDID Testing_Walking_016_016b_000040-a_188191173_rgb
361
+ BIWIRGBDID Testing_Walking_016_016b_000064-a_188193841_rgb
362
+ BIWIRGBDID Testing_Walking_016_016b_000082-a_188195673_rgb
363
+ BIWIRGBDID Testing_Walking_016_016b_000087-a_188196093_rgb
364
+ BIWIRGBDID Testing_Walking_016_016b_000108-a_188198509_rgb
365
+ BIWIRGBDID Testing_Walking_016_016b_000111-a_188198841_rgb
366
+ BIWIRGBDID Testing_Walking_016_016b_000141-a_188202009_rgb
367
+ BIWIRGBDID Testing_Walking_016_016b_000142-a_188202173_rgb
368
+ BIWIRGBDID Testing_Walking_016_016b_000161-a_188204173_rgb
369
+ BIWIRGBDID Testing_Walking_016_016b_000202-a_188208341_rgb
370
+ BIWIRGBDID Testing_Walking_016_016b_000206-a_188208841_rgb
371
+ BIWIRGBDID Testing_Walking_016_016b_000211-a_188209425_rgb
372
+ BIWIRGBDID Testing_Walking_016_016b_000216-a_188209841_rgb
373
+ BIWIRGBDID Testing_Walking_016_016b_000235-a_188211757_rgb
374
+ BIWIRGBDID Testing_Walking_016_016b_000238-a_188212173_rgb
375
+ BIWIRGBDID Testing_Walking_016_016b_000253-a_188213757_rgb
376
+ BIWIRGBDID Testing_Walking_016_016b_000268-a_188215341_rgb
377
+ BIWIRGBDID Testing_Walking_016_016b_000287-a_188217508_rgb
378
+ BIWIRGBDID Testing_Walking_016_016b_000298-a_188218672_rgb
379
+ BIWIRGBDID Testing_Walking_016_016b_000308-a_188219841_rgb
380
+ BIWIRGBDID Testing_Walking_016_016b_000310-a_188220004_rgb
381
+ BIWIRGBDID Testing_Walking_016_016b_000313-a_188220341_rgb
382
+ BIWIRGBDID Testing_Walking_016_016b_000318-a_188220925_rgb
383
+ BIWIRGBDID Testing_Walking_016_016b_000347-a_188224005_rgb
384
+ BIWIRGBDID Testing_Walking_017_017b_000008-a_191574104_rgb
385
+ BIWIRGBDID Testing_Walking_017_017b_000011-a_191574356_rgb
386
+ BIWIRGBDID Testing_Walking_017_017b_000039-a_191577604_rgb
387
+ BIWIRGBDID Testing_Walking_017_017b_000090-a_191583104_rgb
388
+ BIWIRGBDID Testing_Walking_017_017b_000093-a_191583436_rgb
389
+ BIWIRGBDID Testing_Walking_017_017b_000095-a_191583937_rgb
390
+ BIWIRGBDID Testing_Walking_017_017b_000100-a_191584436_rgb
391
+ BIWIRGBDID Testing_Walking_017_017b_000119-a_191586272_rgb
392
+ BIWIRGBDID Testing_Walking_017_017b_000179-a_191592103_rgb
393
+ BIWIRGBDID Testing_Walking_017_017b_000181-a_191592272_rgb
394
+ BIWIRGBDID Testing_Walking_017_017b_000201-a_191594436_rgb
395
+ BIWIRGBDID Testing_Walking_017_017b_000227-a_191597355_rgb
396
+ BIWIRGBDID Testing_Walking_017_017b_000241-a_191598851_rgb
397
+ BIWIRGBDID Testing_Walking_017_017b_000242-a_191598935_rgb
398
+ BIWIRGBDID Testing_Walking_017_017b_000252-a_191599935_rgb
399
+ BIWIRGBDID Testing_Walking_017_017b_000263-a_191600935_rgb
400
+ BIWIRGBDID Testing_Walking_017_017b_000275-a_191602187_rgb
401
+ BIWIRGBDID Testing_Walking_017_017b_000291-a_191603852_rgb
402
+ BIWIRGBDID Testing_Walking_017_017b_000306-a_191605435_rgb
403
+ BIWIRGBDID Testing_Walking_017_017b_000313-a_191606103_rgb
404
+ BIWIRGBDID Testing_Walking_017_017b_000317-a_191606436_rgb
405
+ BIWIRGBDID Testing_Walking_017_017b_000385-a_191613687_rgb
406
+ BIWIRGBDID Testing_Walking_017_017b_000422-a_191617351_rgb
407
+ BIWIRGBDID Testing_Walking_017_017b_000425-a_191617603_rgb
408
+ BIWIRGBDID Testing_Walking_017_017b_000462-a_191622103_rgb
409
+ BIWIRGBDID Testing_Walking_017_017b_000472-a_191623103_rgb
410
+ BIWIRGBDID Testing_Walking_017_017b_000496-a_191625603_rgb
411
+ BIWIRGBDID Testing_Walking_019_019b_000023-a_107227611_rgb
412
+ BIWIRGBDID Testing_Walking_019_019b_000038-a_107228863_rgb
413
+ BIWIRGBDID Testing_Walking_019_019b_000046-a_107229611_rgb
414
+ BIWIRGBDID Testing_Walking_019_019b_000047-a_107229695_rgb
415
+ BIWIRGBDID Testing_Walking_019_019b_000103-a_107235443_rgb
416
+ BIWIRGBDID Testing_Walking_019_019b_000128-a_107238027_rgb
417
+ BIWIRGBDID Testing_Walking_019_019b_000150-a_107240027_rgb
418
+ BIWIRGBDID Testing_Walking_019_019b_000183-a_107243027_rgb
419
+ BIWIRGBDID Testing_Walking_019_019b_000188-a_107243527_rgb
420
+ BIWIRGBDID Testing_Walking_019_019b_000210-a_107245695_rgb
421
+ BIWIRGBDID Testing_Walking_019_019b_000218-a_107246527_rgb
422
+ BIWIRGBDID Testing_Walking_019_019b_000254-a_107250359_rgb
423
+ BIWIRGBDID Testing_Walking_019_019b_000256-a_107250526_rgb
424
+ BIWIRGBDID Testing_Walking_019_019b_000269-a_107251694_rgb
425
+ BIWIRGBDID Testing_Walking_019_019b_000285-a_107253194_rgb
426
+ BIWIRGBDID Testing_Walking_019_019b_000310-a_107255358_rgb
427
+ BIWIRGBDID Testing_Walking_019_019b_000311-a_107255442_rgb
428
+ BIWIRGBDID Testing_Walking_019_019b_000313-a_107255610_rgb
429
+ BIWIRGBDID Testing_Walking_019_019b_000336-a_107258026_rgb
430
+ BIWIRGBDID Testing_Walking_019_019b_000347-a_107259194_rgb
431
+ BIWIRGBDID Testing_Walking_019_019b_000373-a_107261942_rgb
432
+ BIWIRGBDID Testing_Walking_019_019b_000410-a_107265526_rgb
433
+ BIWIRGBDID Testing_Walking_019_019b_000416-a_107266026_rgb
434
+ BIWIRGBDID Testing_Walking_019_019b_000433-a_107267526_rgb
435
+ BIWIRGBDID Testing_Walking_021_021b_000075-a_109053893_rgb
436
+ BIWIRGBDID Testing_Walking_021_021b_000089-a_109055313_rgb
437
+ BIWIRGBDID Testing_Walking_021_021b_000095-a_109055978_rgb
438
+ BIWIRGBDID Testing_Walking_021_021b_000105-a_109056977_rgb
439
+ BIWIRGBDID Testing_Walking_021_021b_000111-a_109057561_rgb
440
+ BIWIRGBDID Testing_Walking_021_021b_000144-a_109060813_rgb
441
+ BIWIRGBDID Testing_Walking_021_021b_000176-a_109063645_rgb
442
+ BIWIRGBDID Testing_Walking_021_021b_000235-a_109069309_rgb
443
+ BIWIRGBDID Testing_Walking_021_021b_000265-a_109072561_rgb
444
+ BIWIRGBDID Testing_Walking_021_021b_000305-a_109076060_rgb
445
+ BIWIRGBDID Testing_Walking_021_021b_000328-a_109078145_rgb
446
+ BIWIRGBDID Testing_Walking_021_021b_000356-a_109080976_rgb
447
+ BIWIRGBDID Testing_Walking_021_021b_000370-a_109082476_rgb
448
+ BIWIRGBDID Testing_Walking_021_021b_000392-a_109084724_rgb
449
+ BIWIRGBDID Testing_Walking_021_021b_000420-a_109087392_rgb
450
+ BIWIRGBDID Testing_Walking_021_021b_000440-a_109089060_rgb
451
+ BIWIRGBDID Testing_Walking_021_021b_000493-a_109093476_rgb
452
+ BIWIRGBDID Testing_Walking_022_022b_000032-a_1214892493_rgb
453
+ BIWIRGBDID Testing_Walking_022_022b_000043-a_1214893409_rgb
454
+ BIWIRGBDID Testing_Walking_022_022b_000052-a_1214894161_rgb
455
+ BIWIRGBDID Testing_Walking_022_022b_000054-a_1214894329_rgb
456
+ BIWIRGBDID Testing_Walking_022_022b_000058-a_1214894661_rgb
457
+ BIWIRGBDID Testing_Walking_022_022b_000087-a_1214897409_rgb
458
+ BIWIRGBDID Testing_Walking_022_022b_000088-a_1214897493_rgb
459
+ BIWIRGBDID Testing_Walking_022_022b_000107-a_1214899409_rgb
460
+ BIWIRGBDID Testing_Walking_022_022b_000126-a_1214901329_rgb
461
+ BIWIRGBDID Testing_Walking_022_022b_000147-a_1214903161_rgb
462
+ BIWIRGBDID Testing_Walking_022_022b_000183-a_1214906493_rgb
463
+ BIWIRGBDID Testing_Walking_022_022b_000185-a_1214906661_rgb
464
+ BIWIRGBDID Testing_Walking_022_022b_000200-a_1214908161_rgb
465
+ BIWIRGBDID Testing_Walking_022_022b_000201-a_1214908245_rgb
466
+ BIWIRGBDID Testing_Walking_022_022b_000206-a_1214908745_rgb
467
+ BIWIRGBDID Testing_Walking_022_022b_000256-a_1214913825_rgb
468
+ BIWIRGBDID Testing_Walking_022_022b_000279-a_1214915909_rgb
469
+ BIWIRGBDID Testing_Walking_022_022b_000288-a_1214916909_rgb
470
+ BIWIRGBDID Testing_Walking_022_022b_000320-a_1214920077_rgb
471
+ BIWIRGBDID Testing_Walking_022_022b_000344-a_1214922660_rgb
472
+ BIWIRGBDID Testing_Walking_022_022b_000360-a_1214924741_rgb
473
+ BIWIRGBDID Testing_Walking_022_022b_000368-a_1214925492_rgb
474
+ BIWIRGBDID Testing_Walking_022_022b_000375-a_1214926076_rgb
475
+ BIWIRGBDID Testing_Walking_022_022b_000379-a_1214926408_rgb
476
+ BIWIRGBDID Testing_Walking_022_022b_000404-a_1214928909_rgb
477
+ BIWIRGBDID Testing_Walking_022_022b_000405-a_1214928992_rgb
478
+ BIWIRGBDID Testing_Walking_022_022b_000437-a_1214931825_rgb
479
+ BIWIRGBDID Testing_Walking_022_022b_000452-a_1214933324_rgb
480
+ BIWIRGBDID Testing_Walking_022_022b_000459-a_1214934076_rgb
481
+ BIWIRGBDID Testing_Walking_023_023b_000009-a_197707271_rgb
482
+ BIWIRGBDID Testing_Walking_023_023b_000020-a_197708439_rgb
483
+ BIWIRGBDID Testing_Walking_023_023b_000029-a_197709270_rgb
484
+ BIWIRGBDID Testing_Walking_023_023b_000051-a_197711186_rgb
485
+ BIWIRGBDID Testing_Walking_023_023b_000063-a_197712438_rgb
486
+ BIWIRGBDID Testing_Walking_023_023b_000077-a_197714271_rgb
487
+ BIWIRGBDID Testing_Walking_023_023b_000085-a_197715354_rgb
488
+ BIWIRGBDID Testing_Walking_023_023b_000099-a_197717186_rgb
489
+ BIWIRGBDID Testing_Walking_023_023b_000112-a_197718854_rgb
490
+ BIWIRGBDID Testing_Walking_023_023b_000118-a_197719518_rgb
491
+ BIWIRGBDID Testing_Walking_023_023b_000119-a_197719602_rgb
492
+ BIWIRGBDID Testing_Walking_023_023b_000124-a_197720186_rgb
493
+ BIWIRGBDID Testing_Walking_023_023b_000130-a_197720686_rgb
494
+ BIWIRGBDID Testing_Walking_023_023b_000155-a_197722935_rgb
495
+ BIWIRGBDID Testing_Walking_023_023b_000181-a_197725602_rgb
496
+ BIWIRGBDID Testing_Walking_023_023b_000197-a_197727855_rgb
497
+ BIWIRGBDID Testing_Walking_023_023b_000211-a_197729770_rgb
498
+ BIWIRGBDID Testing_Walking_023_023b_000222-a_197731018_rgb
499
+ BIWIRGBDID Testing_Walking_023_023b_000233-a_197732186_rgb
500
+ BIWIRGBDID Testing_Walking_023_023b_000253-a_197733850_rgb
501
+ BIWIRGBDID Testing_Walking_023_023b_000261-a_197734602_rgb
502
+ BIWIRGBDID Testing_Walking_023_023b_000266-a_197735018_rgb
503
+ BIWIRGBDID Testing_Walking_023_023b_000276-a_197735934_rgb
504
+ BIWIRGBDID Testing_Walking_023_023b_000279-a_197736185_rgb
505
+ BIWIRGBDID Testing_Walking_023_023b_000285-a_197736686_rgb
506
+ BIWIRGBDID Testing_Walking_023_023b_000307-a_197738850_rgb
507
+ BIWIRGBDID Testing_Walking_023_023b_000310-a_197739186_rgb
508
+ BIWIRGBDID Testing_Walking_023_023b_000328-a_197741518_rgb
509
+ BIWIRGBDID Testing_Walking_023_023b_000356-a_197745433_rgb
510
+ BIWIRGBDID Testing_Walking_023_023b_000371-a_197747017_rgb
511
+ BIWIRGBDID Testing_Walking_023_023b_000379-a_197747685_rgb
512
+ BIWIRGBDID Testing_Walking_023_023b_000407-a_197750265_rgb
513
+ BIWIRGBDID Testing_Walking_023_023b_000416-a_197751018_rgb
514
+ BIWIRGBDID Testing_Walking_023_023b_000424-a_197751765_rgb
515
+ BIWIRGBDID Testing_Walking_023_023b_000438-a_197753517_rgb
516
+ BIWIRGBDID Testing_Walking_024_024b_000028-a_197337876_rgb
517
+ BIWIRGBDID Testing_Walking_024_024b_000035-a_197338544_rgb
518
+ BIWIRGBDID Testing_Walking_024_024b_000058-a_197341708_rgb
519
+ BIWIRGBDID Testing_Walking_024_024b_000061-a_197342128_rgb
520
+ BIWIRGBDID Testing_Walking_024_024b_000065-a_197342627_rgb
521
+ BIWIRGBDID Testing_Walking_024_024b_000074-a_197343628_rgb
522
+ BIWIRGBDID Testing_Walking_024_024b_000090-a_197345628_rgb
523
+ BIWIRGBDID Testing_Walking_024_024b_000132-a_197349628_rgb
524
+ BIWIRGBDID Testing_Walking_024_024b_000184-a_197356127_rgb
525
+ BIWIRGBDID Testing_Walking_024_024b_000210-a_197358459_rgb
526
+ BIWIRGBDID Testing_Walking_024_024b_000239-a_197361544_rgb
527
+ BIWIRGBDID Testing_Walking_024_024b_000259-a_197364043_rgb
528
+ BIWIRGBDID Testing_Walking_024_024b_000304-a_197369043_rgb
529
+ BIWIRGBDID Testing_Walking_024_024b_000355-a_197374707_rgb
530
+ BIWIRGBDID Testing_Walking_026_026b_000007-a_111022853_rgb
531
+ BIWIRGBDID Testing_Walking_026_026b_000025-a_111024354_rgb
532
+ BIWIRGBDID Testing_Walking_026_026b_000038-a_111025770_rgb
533
+ BIWIRGBDID Testing_Walking_026_026b_000061-a_111028185_rgb
534
+ BIWIRGBDID Testing_Walking_026_026b_000062-a_111028269_rgb
535
+ BIWIRGBDID Testing_Walking_026_026b_000067-a_111028769_rgb
536
+ BIWIRGBDID Testing_Walking_026_026b_000072-a_111029353_rgb
537
+ BIWIRGBDID Testing_Walking_026_026b_000083-a_111030521_rgb
538
+ BIWIRGBDID Testing_Walking_026_026b_000097-a_111031769_rgb
539
+ BIWIRGBDID Testing_Walking_026_026b_000102-a_111032186_rgb
540
+ BIWIRGBDID Testing_Walking_026_026b_000108-a_111032854_rgb
541
+ BIWIRGBDID Testing_Walking_026_026b_000114-a_111033354_rgb
542
+ BIWIRGBDID Testing_Walking_026_026b_000136-a_111035605_rgb
543
+ BIWIRGBDID Testing_Walking_026_026b_000160-a_111038185_rgb
544
+ BIWIRGBDID Testing_Walking_026_026b_000162-a_111038353_rgb
545
+ BIWIRGBDID Testing_Walking_026_026b_000179-a_111040105_rgb
546
+ BIWIRGBDID Testing_Walking_026_026b_000213-a_111043769_rgb
547
+ BIWIRGBDID Testing_Walking_026_026b_000215-a_111044021_rgb
548
+ BIWIRGBDID Testing_Walking_026_026b_000221-a_111044521_rgb
549
+ BIWIRGBDID Testing_Walking_026_026b_000232-a_111045602_rgb
550
+ BIWIRGBDID Testing_Walking_026_026b_000247-a_111047185_rgb
551
+ BIWIRGBDID Testing_Walking_026_026b_000249-a_111047353_rgb
552
+ BIWIRGBDID Testing_Walking_026_026b_000265-a_111049021_rgb
553
+ BIWIRGBDID Testing_Walking_026_026b_000275-a_111050101_rgb
554
+ BIWIRGBDID Testing_Walking_026_026b_000333-a_111055601_rgb
555
+ BIWIRGBDID Testing_Walking_027_027b_000001-a_111418663_rgb
556
+ BIWIRGBDID Testing_Walking_027_027b_000017-a_111419995_rgb
557
+ BIWIRGBDID Testing_Walking_027_027b_000028-a_111420911_rgb
558
+ BIWIRGBDID Testing_Walking_027_027b_000049-a_111422912_rgb
559
+ BIWIRGBDID Testing_Walking_027_027b_000062-a_111424247_rgb
560
+ BIWIRGBDID Testing_Walking_027_027b_000072-a_111425327_rgb
561
+ BIWIRGBDID Testing_Walking_027_027b_000080-a_111426163_rgb
562
+ BIWIRGBDID Testing_Walking_027_027b_000088-a_111426995_rgb
563
+ BIWIRGBDID Testing_Walking_027_027b_000098-a_111428079_rgb
564
+ BIWIRGBDID Testing_Walking_027_027b_000132-a_111431163_rgb
565
+ BIWIRGBDID Testing_Walking_027_027b_000198-a_111436827_rgb
566
+ BIWIRGBDID Testing_Walking_027_027b_000203-a_111437412_rgb
567
+ BIWIRGBDID Testing_Walking_027_027b_000205-a_111437579_rgb
568
+ BIWIRGBDID Testing_Walking_027_027b_000207-a_111437827_rgb
569
+ BIWIRGBDID Testing_Walking_027_027b_000218-a_111438911_rgb
570
+ BIWIRGBDID Testing_Walking_027_027b_000240-a_111441411_rgb
571
+ BIWIRGBDID Testing_Walking_027_027b_000313-a_111448663_rgb
572
+ BIWIRGBDID Testing_Walking_027_027b_000323-a_111449663_rgb
573
+ BIWIRGBDID Testing_Walking_027_027b_000338-a_111451243_rgb
574
+ BIWIRGBDID Testing_Walking_027_027b_000350-a_111452495_rgb
575
+ BIWIRGBDID Testing_Walking_027_027b_000373-a_111454911_rgb
576
+ BIWIRGBDID Testing_Walking_027_027b_000380-a_111455495_rgb
577
+ BIWIRGBDID Testing_Walking_027_027b_000382-a_111455659_rgb
578
+ BIWIRGBDID Testing_Walking_027_027b_000405-a_111457743_rgb
579
+ BIWIRGBDID Testing_Walking_027_027b_000412-a_111458495_rgb
580
+ BIWIRGBDID Testing_Walking_030_030b_000015-a_1233809555_rgb
581
+ BIWIRGBDID Testing_Walking_030_030b_000022-a_1233810139_rgb
582
+ BIWIRGBDID Testing_Walking_030_030b_000068-a_1233815139_rgb
583
+ BIWIRGBDID Testing_Walking_030_030b_000075-a_1233815970_rgb
584
+ BIWIRGBDID Testing_Walking_030_030b_000077-a_1233816139_rgb
585
+ BIWIRGBDID Testing_Walking_030_030b_000079-a_1233816307_rgb
586
+ BIWIRGBDID Testing_Walking_030_030b_000102-a_1233818638_rgb
587
+ BIWIRGBDID Testing_Walking_030_030b_000129-a_1233820886_rgb
588
+ BIWIRGBDID Testing_Walking_030_030b_000156-a_1233823886_rgb
589
+ BIWIRGBDID Testing_Walking_030_030b_000168-a_1233825054_rgb
590
+ BIWIRGBDID Testing_Walking_030_030b_000182-a_1233826638_rgb
591
+ BIWIRGBDID Testing_Walking_030_030b_000197-a_1233828222_rgb
592
+ BIWIRGBDID Testing_Walking_030_030b_000216-a_1233830138_rgb
593
+ BIWIRGBDID Testing_Walking_030_030b_000219-a_1233830386_rgb
594
+ BIWIRGBDID Testing_Walking_030_030b_000223-a_1233830722_rgb
595
+ BIWIRGBDID Testing_Walking_030_030b_000226-a_1233830970_rgb
596
+ BIWIRGBDID Testing_Walking_030_030b_000234-a_1233831722_rgb
597
+ BIWIRGBDID Testing_Walking_030_030b_000248-a_1233832886_rgb
598
+ BIWIRGBDID Testing_Walking_030_030b_000299-a_1233838054_rgb
599
+ BIWIRGBDID Testing_Walking_030_030b_000334-a_1233841722_rgb
600
+ BIWIRGBDID Testing_Walking_030_030b_000344-a_1233842554_rgb
601
+ BIWIRGBDID Testing_Walking_030_030b_000350-a_1233843218_rgb
602
+ BIWIRGBDID Testing_Walking_030_030b_000351-a_1233843386_rgb
603
+ BIWIRGBDID Testing_Walking_030_030b_000357-a_1233844302_rgb
604
+ BIWIRGBDID Testing_Walking_030_030b_000367-a_1233845302_rgb
605
+ BIWIRGBDID Testing_Walking_030_030b_000398-a_1233848638_rgb
606
+ BIWIRGBDID Testing_Walking_031_031b_000001-a_186803438_rgb
607
+ BIWIRGBDID Testing_Walking_031_031b_000005-a_186803770_rgb
608
+ BIWIRGBDID Testing_Walking_031_031b_000054-a_186808770_rgb
609
+ BIWIRGBDID Testing_Walking_031_031b_000069-a_186810354_rgb
610
+ BIWIRGBDID Testing_Walking_031_031b_000079-a_186811438_rgb
611
+ BIWIRGBDID Testing_Walking_031_031b_000125-a_186816438_rgb
612
+ BIWIRGBDID Testing_Walking_031_031b_000132-a_186817102_rgb
613
+ BIWIRGBDID Testing_Walking_031_031b_000165-a_186820434_rgb
614
+ BIWIRGBDID Testing_Walking_031_031b_000176-a_186821518_rgb
615
+ BIWIRGBDID Testing_Walking_031_031b_000197-a_186823770_rgb
616
+ BIWIRGBDID Testing_Walking_031_031b_000215-a_186825601_rgb
617
+ BIWIRGBDID Testing_Walking_031_031b_000221-a_186826102_rgb
618
+ BIWIRGBDID Testing_Walking_031_031b_000245-a_186828602_rgb
619
+ BIWIRGBDID Testing_Walking_031_031b_000259-a_186829934_rgb
620
+ BIWIRGBDID Testing_Walking_031_031b_000269-a_186831018_rgb
621
+ BIWIRGBDID Testing_Walking_031_031b_000287-a_186832933_rgb
622
+ BIWIRGBDID Testing_Walking_031_031b_000341-a_186838350_rgb
623
+ BIWIRGBDID Testing_Walking_031_031b_000353-a_186839350_rgb
624
+ BIWIRGBDID Testing_Walking_031_031b_000361-a_186840101_rgb
625
+ BIWIRGBDID Testing_Walking_031_031b_000368-a_186840685_rgb
626
+ BIWIRGBDID Testing_Walking_031_031b_000371-a_186841017_rgb
627
+ BIWIRGBDID Testing_Walking_031_031b_000380-a_186841933_rgb
628
+ BIWIRGBDID Testing_Walking_031_031b_000400-a_186844101_rgb
629
+ BIWIRGBDID Testing_Walking_031_031b_000402-a_186844349_rgb
630
+ BIWIRGBDID Testing_Walking_031_031b_000403-a_186844433_rgb
631
+ BIWIRGBDID Testing_Walking_032_032b_000022-a_111090932_rgb
632
+ BIWIRGBDID Testing_Walking_032_032b_000034-a_111092016_rgb
633
+ BIWIRGBDID Testing_Walking_032_032b_000038-a_111092516_rgb
634
+ BIWIRGBDID Testing_Walking_032_032b_000044-a_111093100_rgb
635
+ BIWIRGBDID Testing_Walking_032_032b_000053-a_111094100_rgb
636
+ BIWIRGBDID Testing_Walking_032_032b_000127-a_111101515_rgb
637
+ BIWIRGBDID Testing_Walking_032_032b_000128-a_111101599_rgb
638
+ BIWIRGBDID Testing_Walking_032_032b_000154-a_111104015_rgb
639
+ BIWIRGBDID Testing_Walking_032_032b_000160-a_111104684_rgb
640
+ BIWIRGBDID Testing_Walking_032_032b_000184-a_111107100_rgb
641
+ BIWIRGBDID Testing_Walking_032_032b_000187-a_111107515_rgb
642
+ BIWIRGBDID Testing_Walking_032_032b_000254-a_111114099_rgb
643
+ BIWIRGBDID Testing_Walking_032_032b_000302-a_111118431_rgb
644
+ BIWIRGBDID Testing_Walking_032_032b_000308-a_111119015_rgb
645
+ BIWIRGBDID Testing_Walking_032_032b_000313-a_111119599_rgb
646
+ BIWIRGBDID Testing_Walking_032_032b_000324-a_111120763_rgb
647
+ BIWIRGBDID Testing_Walking_032_032b_000332-a_111121683_rgb
648
+ BIWIRGBDID Testing_Walking_032_032b_000335-a_111122015_rgb
649
+ BIWIRGBDID Testing_Walking_032_032b_000351-a_111123683_rgb
650
+ BIWIRGBDID Testing_Walking_032_032b_000363-a_111124763_rgb
651
+ BIWIRGBDID Testing_Walking_032_032b_000385-a_111126847_rgb
652
+ BIWIRGBDID Testing_Walking_032_032b_000404-a_111128763_rgb
653
+ BIWIRGBDID Testing_Walking_036_036b_000040-a_107787409_rgb
654
+ BIWIRGBDID Testing_Walking_036_036b_000049-a_107788409_rgb
655
+ BIWIRGBDID Testing_Walking_036_036b_000121-a_107795493_rgb
656
+ BIWIRGBDID Testing_Walking_036_036b_000150-a_107798077_rgb
657
+ BIWIRGBDID Testing_Walking_036_036b_000152-a_107798241_rgb
658
+ BIWIRGBDID Testing_Walking_036_036b_000176-a_107800661_rgb
659
+ BIWIRGBDID Testing_Walking_036_036b_000181-a_107801241_rgb
660
+ BIWIRGBDID Testing_Walking_036_036b_000196-a_107802825_rgb
661
+ BIWIRGBDID Testing_Walking_036_036b_000210-a_107804241_rgb
662
+ BIWIRGBDID Testing_Walking_036_036b_000213-a_107804577_rgb
663
+ BIWIRGBDID Testing_Walking_036_036b_000215-a_107804824_rgb
664
+ BIWIRGBDID Testing_Walking_036_036b_000220-a_107805325_rgb
665
+ BIWIRGBDID Testing_Walking_036_036b_000226-a_107805908_rgb
666
+ BIWIRGBDID Testing_Walking_036_036b_000253-a_107808324_rgb
667
+ BIWIRGBDID Testing_Walking_036_036b_000277-a_107810657_rgb
668
+ BIWIRGBDID Testing_Walking_036_036b_000292-a_107812240_rgb
669
+ BIWIRGBDID Testing_Walking_036_036b_000303-a_107813324_rgb
670
+ BIWIRGBDID Testing_Walking_036_036b_000344-a_107817741_rgb
671
+ BIWIRGBDID Testing_Walking_036_036b_000373-a_107820408_rgb
672
+ BIWIRGBDID Testing_Walking_036_036b_000387-a_107821656_rgb
673
+ BIWIRGBDID Testing_Walking_037_037b_000029-a_191452363_rgb
674
+ BIWIRGBDID Testing_Walking_037_037b_000054-a_191455364_rgb
675
+ BIWIRGBDID Testing_Walking_037_037b_000058-a_191455779_rgb
676
+ BIWIRGBDID Testing_Walking_037_037b_000075-a_191457527_rgb
677
+ BIWIRGBDID Testing_Walking_037_037b_000127-a_191463363_rgb
678
+ BIWIRGBDID Testing_Walking_037_037b_000146-a_191465279_rgb
679
+ BIWIRGBDID Testing_Walking_037_037b_000166-a_191467527_rgb
680
+ BIWIRGBDID Testing_Walking_037_037b_000169-a_191467863_rgb
681
+ BIWIRGBDID Testing_Walking_037_037b_000179-a_191469027_rgb
682
+ BIWIRGBDID Testing_Walking_037_037b_000188-a_191469947_rgb
683
+ BIWIRGBDID Testing_Walking_037_037b_000215-a_191473110_rgb
684
+ BIWIRGBDID Testing_Walking_037_037b_000227-a_191474362_rgb
685
+ BIWIRGBDID Testing_Walking_037_037b_000251-a_191476694_rgb
686
+ BIWIRGBDID Testing_Walking_037_037b_000254-a_191476942_rgb
687
+ BIWIRGBDID Testing_Walking_037_037b_000259-a_191477362_rgb
688
+ BIWIRGBDID Testing_Walking_037_037b_000270-a_191478694_rgb
689
+ BIWIRGBDID Testing_Walking_037_037b_000277-a_191479363_rgb
690
+ BIWIRGBDID Testing_Walking_037_037b_000291-a_191480695_rgb
691
+ BIWIRGBDID Testing_Walking_037_037b_000299-a_191481694_rgb
692
+ BIWIRGBDID Testing_Walking_037_037b_000308-a_191482778_rgb
693
+ BIWIRGBDID Testing_Walking_037_037b_000322-a_191484194_rgb
694
+ BIWIRGBDID Testing_Walking_037_037b_000323-a_191484362_rgb
695
+ BIWIRGBDID Testing_Walking_037_037b_000345-a_191486610_rgb
696
+ BIWIRGBDID Testing_Walking_039_039b_000002-a_1219323133_rgb
697
+ BIWIRGBDID Testing_Walking_039_039b_000017-a_1219324469_rgb
698
+ BIWIRGBDID Testing_Walking_039_039b_000033-a_1219326048_rgb
699
+ BIWIRGBDID Testing_Walking_039_039b_000075-a_1219330465_rgb
700
+ BIWIRGBDID Testing_Walking_039_039b_000083-a_1219331300_rgb
701
+ BIWIRGBDID Testing_Walking_039_039b_000109-a_1219333632_rgb
702
+ BIWIRGBDID Testing_Walking_039_039b_000123-a_1219334800_rgb
703
+ BIWIRGBDID Testing_Walking_039_039b_000126-a_1219335048_rgb
704
+ BIWIRGBDID Testing_Walking_039_039b_000136-a_1219335884_rgb
705
+ BIWIRGBDID Testing_Walking_039_039b_000150-a_1219337300_rgb
706
+ BIWIRGBDID Testing_Walking_039_039b_000158-a_1219338132_rgb
707
+ BIWIRGBDID Testing_Walking_039_039b_000234-a_1219345216_rgb
708
+ BIWIRGBDID Testing_Walking_039_039b_000253-a_1219346964_rgb
709
+ BIWIRGBDID Testing_Walking_039_039b_000259-a_1219347632_rgb
710
+ BIWIRGBDID Testing_Walking_039_039b_000261-a_1219347880_rgb
711
+ BIWIRGBDID Testing_Walking_039_039b_000274-a_1219349300_rgb
712
+ BIWIRGBDID Testing_Walking_039_039b_000286-a_1219350548_rgb
713
+ BIWIRGBDID Testing_Walking_039_039b_000294-a_1219351380_rgb
714
+ BIWIRGBDID Testing_Walking_039_039b_000295-a_1219351464_rgb
715
+ BIWIRGBDID Testing_Walking_039_039b_000302-a_1219352216_rgb
716
+ BIWIRGBDID Testing_Walking_039_039b_000304-a_1219352380_rgb
717
+ BIWIRGBDID Testing_Walking_039_039b_000305-a_1219352464_rgb
718
+ BIWIRGBDID Testing_Walking_039_039b_000327-a_1219354300_rgb
719
+ BIWIRGBDID Testing_Walking_039_039b_000336-a_1219355048_rgb
720
+ BIWIRGBDID Testing_Walking_039_039b_000375-a_1219358632_rgb
721
+ BIWIRGBDID Testing_Walking_040_040b_000034-a_1231749844_rgb
722
+ BIWIRGBDID Testing_Walking_040_040b_000047-a_1231751260_rgb
723
+ BIWIRGBDID Testing_Walking_040_040b_000063-a_1231752924_rgb
724
+ BIWIRGBDID Testing_Walking_040_040b_000074-a_1231754176_rgb
725
+ BIWIRGBDID Testing_Walking_040_040b_000087-a_1231755508_rgb
726
+ BIWIRGBDID Testing_Walking_040_040b_000121-a_1231759260_rgb
727
+ BIWIRGBDID Testing_Walking_040_040b_000127-a_1231759840_rgb
728
+ BIWIRGBDID Testing_Walking_040_040b_000129-a_1231760092_rgb
729
+ BIWIRGBDID Testing_Walking_040_040b_000138-a_1231761260_rgb
730
+ BIWIRGBDID Testing_Walking_040_040b_000148-a_1231762340_rgb
731
+ BIWIRGBDID Testing_Walking_040_040b_000184-a_1231766260_rgb
732
+ BIWIRGBDID Testing_Walking_040_040b_000250-a_1231773424_rgb
733
+ BIWIRGBDID Testing_Walking_040_040b_000308-a_1231779840_rgb
734
+ BIWIRGBDID Testing_Walking_040_040b_000384-a_1231787923_rgb
735
+ BIWIRGBDID Testing_Walking_044_044b_000036-a_1214190213_rgb
736
+ BIWIRGBDID Testing_Walking_044_044b_000105-a_1214196713_rgb
737
+ BIWIRGBDID Testing_Walking_044_044b_000108-a_1214197045_rgb
738
+ BIWIRGBDID Testing_Walking_044_044b_000134-a_1214199797_rgb
739
+ BIWIRGBDID Testing_Walking_044_044b_000152-a_1214201381_rgb
740
+ BIWIRGBDID Testing_Walking_044_044b_000161-a_1214202213_rgb
741
+ BIWIRGBDID Testing_Walking_044_044b_000262-a_1214211712_rgb
742
+ BIWIRGBDID Testing_Walking_044_044b_000266-a_1214212129_rgb
743
+ BIWIRGBDID Testing_Walking_044_044b_000285-a_1214214128_rgb
744
+ BIWIRGBDID Testing_Walking_044_044b_000309-a_1214216544_rgb
745
+ BIWIRGBDID Testing_Walking_044_044b_000348-a_1214220128_rgb
746
+ BIWIRGBDID Testing_Walking_044_044b_000354-a_1214220712_rgb
747
+ BIWIRGBDID Testing_Walking_046_046b_000002-a_111236092_rgb
748
+ BIWIRGBDID Testing_Walking_046_046b_000023-a_111238008_rgb
749
+ BIWIRGBDID Testing_Walking_046_046b_000032-a_111238840_rgb
750
+ BIWIRGBDID Testing_Walking_046_046b_000057-a_111241172_rgb
751
+ BIWIRGBDID Testing_Walking_046_046b_000080-a_111243672_rgb
752
+ BIWIRGBDID Testing_Walking_046_046b_000103-a_111246256_rgb
753
+ BIWIRGBDID Testing_Walking_046_046b_000120-a_111247756_rgb
754
+ BIWIRGBDID Testing_Walking_046_046b_000128-a_111248508_rgb
755
+ BIWIRGBDID Testing_Walking_046_046b_000168-a_111252508_rgb
756
+ BIWIRGBDID Testing_Walking_046_046b_000171-a_111252840_rgb
757
+ BIWIRGBDID Testing_Walking_046_046b_000242-a_111259588_rgb
758
+ BIWIRGBDID Testing_Walking_046_046b_000281-a_111263672_rgb
759
+ BIWIRGBDID Testing_Walking_046_046b_000323-a_111267587_rgb
760
+ BIWIRGBDID Testing_Walking_046_046b_000335-a_111268755_rgb
761
+ BIWIRGBDID Testing_Walking_046_046b_000342-a_111269507_rgb
762
+ BIWIRGBDID Testing_Walking_046_046b_000344-a_111269756_rgb
763
+ BIWIRGBDID Testing_Walking_047_047b_000014-a_1214112719_rgb
764
+ BIWIRGBDID Testing_Walking_047_047b_000024-a_1214113551_rgb
765
+ BIWIRGBDID Testing_Walking_047_047b_000035-a_1214114635_rgb
766
+ BIWIRGBDID Testing_Walking_047_047b_000070-a_1214118467_rgb
767
+ BIWIRGBDID Testing_Walking_047_047b_000083-a_1214119719_rgb
768
+ BIWIRGBDID Testing_Walking_047_047b_000094-a_1214120719_rgb
769
+ BIWIRGBDID Testing_Walking_047_047b_000105-a_1214122135_rgb
770
+ BIWIRGBDID Testing_Walking_047_047b_000128-a_1214124635_rgb
771
+ BIWIRGBDID Testing_Walking_047_047b_000144-a_1214126551_rgb
772
+ BIWIRGBDID Testing_Walking_047_047b_000146-a_1214126718_rgb
773
+ BIWIRGBDID Testing_Walking_047_047b_000166-a_1214128634_rgb
774
+ BIWIRGBDID Testing_Walking_047_047b_000186-a_1214130303_rgb
775
+ BIWIRGBDID Testing_Walking_047_047b_000197-a_1214131298_rgb
776
+ BIWIRGBDID Testing_Walking_047_047b_000207-a_1214132218_rgb
777
+ BIWIRGBDID Testing_Walking_047_047b_000209-a_1214132382_rgb
778
+ BIWIRGBDID Testing_Walking_047_047b_000213-a_1214132718_rgb
779
+ BIWIRGBDID Testing_Walking_047_047b_000228-a_1214134299_rgb
780
+ BIWIRGBDID Testing_Walking_047_047b_000261-a_1214138134_rgb
781
+ BIWIRGBDID Testing_Walking_047_047b_000303-a_1214142218_rgb
782
+ BIWIRGBDID Testing_Walking_047_047b_000308-a_1214142882_rgb
783
+ BIWIRGBDID Testing_Walking_047_047b_000314-a_1214143466_rgb
784
+ BIWIRGBDID Testing_Walking_047_047b_000336-a_1214145718_rgb
785
+ BIWIRGBDID Testing_Walking_049_049b_000002-a_201380445_rgb
786
+ BIWIRGBDID Testing_Walking_049_049b_000132-a_201394025_rgb
787
+ BIWIRGBDID Testing_Walking_049_049b_000138-a_201394693_rgb
788
+ BIWIRGBDID Testing_Walking_049_049b_000145-a_201395360_rgb
789
+ BIWIRGBDID Testing_Walking_049_049b_000166-a_201397609_rgb
790
+ BIWIRGBDID Testing_Walking_049_049b_000257-a_201407193_rgb
791
+ BIWIRGBDID Testing_Walking_049_049b_000272-a_201408776_rgb
792
+ BIWIRGBDID Testing_Walking_049_049b_000287-a_201410440_rgb
793
+ BIWIRGBDID Testing_Walking_049_049b_000295-a_201411276_rgb
794
+ BIWIRGBDID Testing_Walking_049_049b_000312-a_201413192_rgb
795
+ BIWIRGBDID Testing_Walking_049_049b_000316-a_201413608_rgb
796
+ BIWIRGBDID Testing_Walking_049_049b_000335-a_201415608_rgb
797
+ BIWIRGBDID Testing_Walking_049_049b_000366-a_201418940_rgb
798
+ BIWIRGBDID Testing_Walking_049_049b_000395-a_201422108_rgb
799
+ BIWIRGBDID Testing_Walking_049_049b_000409-a_201423524_rgb
800
+ BIWIRGBDID Testing_Walking_049_049b_000429-a_201425692_rgb
external/WildCamera/splits/cad120_test.txt ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 1.jpg
2
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 11.jpg
3
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 61.jpg
4
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 71.jpg
5
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 121.jpg
6
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 131.jpg
7
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 141.jpg
8
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 151.jpg
9
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 161.jpg
10
+ CAD120_Subject1_rgbd_images_arranging_objects_sequence 171.jpg
11
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 1.jpg
12
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 41.jpg
13
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 61.jpg
14
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 71.jpg
15
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 81.jpg
16
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 101.jpg
17
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 111.jpg
18
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 121.jpg
19
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 131.jpg
20
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 141.jpg
21
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 151.jpg
22
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 161.jpg
23
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 171.jpg
24
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 181.jpg
25
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 201.jpg
26
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 221.jpg
27
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 241.jpg
28
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 291.jpg
29
+ CAD120_Subject1_rgbd_images_cleaning_objects_sequence 321.jpg
30
+ CAD120_Subject1_rgbd_images_having_meal_sequence 1.jpg
31
+ CAD120_Subject1_rgbd_images_having_meal_sequence 71.jpg
32
+ CAD120_Subject1_rgbd_images_having_meal_sequence 81.jpg
33
+ CAD120_Subject1_rgbd_images_having_meal_sequence 101.jpg
34
+ CAD120_Subject1_rgbd_images_having_meal_sequence 141.jpg
35
+ CAD120_Subject1_rgbd_images_having_meal_sequence 161.jpg
36
+ CAD120_Subject1_rgbd_images_having_meal_sequence 171.jpg
37
+ CAD120_Subject1_rgbd_images_having_meal_sequence 201.jpg
38
+ CAD120_Subject1_rgbd_images_having_meal_sequence 211.jpg
39
+ CAD120_Subject1_rgbd_images_having_meal_sequence 221.jpg
40
+ CAD120_Subject1_rgbd_images_having_meal_sequence 231.jpg
41
+ CAD120_Subject1_rgbd_images_having_meal_sequence 241.jpg
42
+ CAD120_Subject1_rgbd_images_having_meal_sequence 281.jpg
43
+ CAD120_Subject1_rgbd_images_having_meal_sequence 291.jpg
44
+ CAD120_Subject1_rgbd_images_having_meal_sequence 311.jpg
45
+ CAD120_Subject1_rgbd_images_having_meal_sequence 331.jpg
46
+ CAD120_Subject1_rgbd_images_having_meal_sequence 341.jpg
47
+ CAD120_Subject1_rgbd_images_having_meal_sequence 361.jpg
48
+ CAD120_Subject1_rgbd_images_having_meal_sequence 381.jpg
49
+ CAD120_Subject1_rgbd_images_having_meal_sequence 391.jpg
50
+ CAD120_Subject1_rgbd_images_having_meal_sequence 401.jpg
51
+ CAD120_Subject1_rgbd_images_having_meal_sequence 411.jpg
52
+ CAD120_Subject1_rgbd_images_having_meal_sequence 421.jpg
53
+ CAD120_Subject1_rgbd_images_having_meal_sequence 431.jpg
54
+ CAD120_Subject1_rgbd_images_having_meal_sequence 441.jpg
55
+ CAD120_Subject1_rgbd_images_having_meal_sequence 471.jpg
56
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 1.jpg
57
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 11.jpg
58
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 41.jpg
59
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 51.jpg
60
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 81.jpg
61
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 91.jpg
62
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 101.jpg
63
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 111.jpg
64
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 131.jpg
65
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 151.jpg
66
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 161.jpg
67
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 171.jpg
68
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 211.jpg
69
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 221.jpg
70
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 231.jpg
71
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 261.jpg
72
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 281.jpg
73
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 301.jpg
74
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 321.jpg
75
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 351.jpg
76
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 371.jpg
77
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 381.jpg
78
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 401.jpg
79
+ CAD120_Subject1_rgbd_images_making_cereal_sequence 421.jpg
80
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 1.jpg
81
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 11.jpg
82
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 21.jpg
83
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 31.jpg
84
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 41.jpg
85
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 91.jpg
86
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 101.jpg
87
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 111.jpg
88
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 121.jpg
89
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 131.jpg
90
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 141.jpg
91
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 191.jpg
92
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 201.jpg
93
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 221.jpg
94
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 231.jpg
95
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 241.jpg
96
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 261.jpg
97
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 281.jpg
98
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 291.jpg
99
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 301.jpg
100
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 321.jpg
101
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 341.jpg
102
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 351.jpg
103
+ CAD120_Subject1_rgbd_images_microwaving_food_sequence 381.jpg
104
+ CAD120_Subject1_rgbd_images_picking_objects_sequence 1.jpg
105
+ CAD120_Subject1_rgbd_images_picking_objects_sequence 31.jpg
106
+ CAD120_Subject1_rgbd_images_picking_objects_sequence 41.jpg
107
+ CAD120_Subject1_rgbd_images_picking_objects_sequence 71.jpg
108
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 1.jpg
109
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 11.jpg
110
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 21.jpg
111
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 31.jpg
112
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 51.jpg
113
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 61.jpg
114
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 71.jpg
115
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 91.jpg
116
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 101.jpg
117
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 121.jpg
118
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 131.jpg
119
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 141.jpg
120
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 161.jpg
121
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 171.jpg
122
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 211.jpg
123
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 221.jpg
124
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 231.jpg
125
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 241.jpg
126
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 251.jpg
127
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 271.jpg
128
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 281.jpg
129
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 301.jpg
130
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 321.jpg
131
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 341.jpg
132
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 351.jpg
133
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 361.jpg
134
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 371.jpg
135
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 381.jpg
136
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 401.jpg
137
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 411.jpg
138
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 421.jpg
139
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 441.jpg
140
+ CAD120_Subject1_rgbd_images_stacking_objects_sequence 461.jpg
141
+ CAD120_Subject1_rgbd_images_taking_food_sequence 11.jpg
142
+ CAD120_Subject1_rgbd_images_taking_food_sequence 21.jpg
143
+ CAD120_Subject1_rgbd_images_taking_food_sequence 31.jpg
144
+ CAD120_Subject1_rgbd_images_taking_food_sequence 41.jpg
145
+ CAD120_Subject1_rgbd_images_taking_food_sequence 71.jpg
146
+ CAD120_Subject1_rgbd_images_taking_food_sequence 81.jpg
147
+ CAD120_Subject1_rgbd_images_taking_food_sequence 91.jpg
148
+ CAD120_Subject1_rgbd_images_taking_food_sequence 111.jpg
149
+ CAD120_Subject1_rgbd_images_taking_food_sequence 131.jpg
150
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 21.jpg
151
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 31.jpg
152
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 41.jpg
153
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 51.jpg
154
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 71.jpg
155
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 81.jpg
156
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 91.jpg
157
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 101.jpg
158
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 111.jpg
159
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 121.jpg
160
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 131.jpg
161
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 141.jpg
162
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 151.jpg
163
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 171.jpg
164
+ CAD120_Subject1_rgbd_images_taking_medicine_sequence 201.jpg
165
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 1.jpg
166
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 11.jpg
167
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 51.jpg
168
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 61.jpg
169
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 71.jpg
170
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 81.jpg
171
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 111.jpg
172
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 141.jpg
173
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 151.jpg
174
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 161.jpg
175
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 181.jpg
176
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 201.jpg
177
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 221.jpg
178
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 231.jpg
179
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 241.jpg
180
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 261.jpg
181
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 291.jpg
182
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 311.jpg
183
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 351.jpg
184
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 361.jpg
185
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 371.jpg
186
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 391.jpg
187
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 401.jpg
188
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 461.jpg
189
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 471.jpg
190
+ CAD120_Subject1_rgbd_images_unstacking_objects_sequence 501.jpg
191
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 1.jpg
192
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 11.jpg
193
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 21.jpg
194
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 31.jpg
195
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 41.jpg
196
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 51.jpg
197
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 61.jpg
198
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 71.jpg
199
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 81.jpg
200
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 101.jpg
201
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 111.jpg
202
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 131.jpg
203
+ CAD120_Subject3_rgbd_images_arranging_objects_sequence 141.jpg
204
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 1.jpg
205
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 21.jpg
206
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 31.jpg
207
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 41.jpg
208
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 61.jpg
209
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 71.jpg
210
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 101.jpg
211
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 141.jpg
212
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 161.jpg
213
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 181.jpg
214
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 191.jpg
215
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 211.jpg
216
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 221.jpg
217
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 241.jpg
218
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 251.jpg
219
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 261.jpg
220
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 271.jpg
221
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 281.jpg
222
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 291.jpg
223
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 311.jpg
224
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 331.jpg
225
+ CAD120_Subject3_rgbd_images_cleaning_objects_sequence 341.jpg
226
+ CAD120_Subject3_rgbd_images_having_meal_sequence 1.jpg
227
+ CAD120_Subject3_rgbd_images_having_meal_sequence 11.jpg
228
+ CAD120_Subject3_rgbd_images_having_meal_sequence 21.jpg
229
+ CAD120_Subject3_rgbd_images_having_meal_sequence 51.jpg
230
+ CAD120_Subject3_rgbd_images_having_meal_sequence 61.jpg
231
+ CAD120_Subject3_rgbd_images_having_meal_sequence 91.jpg
232
+ CAD120_Subject3_rgbd_images_having_meal_sequence 101.jpg
233
+ CAD120_Subject3_rgbd_images_having_meal_sequence 111.jpg
234
+ CAD120_Subject3_rgbd_images_having_meal_sequence 121.jpg
235
+ CAD120_Subject3_rgbd_images_having_meal_sequence 131.jpg
236
+ CAD120_Subject3_rgbd_images_having_meal_sequence 151.jpg
237
+ CAD120_Subject3_rgbd_images_having_meal_sequence 201.jpg
238
+ CAD120_Subject3_rgbd_images_having_meal_sequence 221.jpg
239
+ CAD120_Subject3_rgbd_images_having_meal_sequence 241.jpg
240
+ CAD120_Subject3_rgbd_images_having_meal_sequence 261.jpg
241
+ CAD120_Subject3_rgbd_images_having_meal_sequence 271.jpg
242
+ CAD120_Subject3_rgbd_images_having_meal_sequence 301.jpg
243
+ CAD120_Subject3_rgbd_images_having_meal_sequence 321.jpg
244
+ CAD120_Subject3_rgbd_images_having_meal_sequence 341.jpg
245
+ CAD120_Subject3_rgbd_images_having_meal_sequence 371.jpg
246
+ CAD120_Subject3_rgbd_images_having_meal_sequence 381.jpg
247
+ CAD120_Subject3_rgbd_images_having_meal_sequence 411.jpg
248
+ CAD120_Subject3_rgbd_images_having_meal_sequence 421.jpg
249
+ CAD120_Subject3_rgbd_images_having_meal_sequence 441.jpg
250
+ CAD120_Subject3_rgbd_images_having_meal_sequence 461.jpg
251
+ CAD120_Subject3_rgbd_images_having_meal_sequence 471.jpg
252
+ CAD120_Subject3_rgbd_images_having_meal_sequence 481.jpg
253
+ CAD120_Subject3_rgbd_images_having_meal_sequence 491.jpg
254
+ CAD120_Subject3_rgbd_images_having_meal_sequence 501.jpg
255
+ CAD120_Subject3_rgbd_images_having_meal_sequence 521.jpg
256
+ CAD120_Subject3_rgbd_images_having_meal_sequence 531.jpg
257
+ CAD120_Subject3_rgbd_images_having_meal_sequence 551.jpg
258
+ CAD120_Subject3_rgbd_images_having_meal_sequence 611.jpg
259
+ CAD120_Subject3_rgbd_images_having_meal_sequence 621.jpg
260
+ CAD120_Subject3_rgbd_images_having_meal_sequence 651.jpg
261
+ CAD120_Subject3_rgbd_images_having_meal_sequence 661.jpg
262
+ CAD120_Subject3_rgbd_images_having_meal_sequence 671.jpg
263
+ CAD120_Subject3_rgbd_images_having_meal_sequence 691.jpg
264
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 41.jpg
265
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 51.jpg
266
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 61.jpg
267
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 71.jpg
268
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 111.jpg
269
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 141.jpg
270
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 151.jpg
271
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 161.jpg
272
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 171.jpg
273
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 211.jpg
274
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 231.jpg
275
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 241.jpg
276
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 291.jpg
277
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 301.jpg
278
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 321.jpg
279
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 341.jpg
280
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 351.jpg
281
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 361.jpg
282
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 371.jpg
283
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 411.jpg
284
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 421.jpg
285
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 431.jpg
286
+ CAD120_Subject3_rgbd_images_making_cereal_sequence 441.jpg
287
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 1.jpg
288
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 11.jpg
289
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 41.jpg
290
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 51.jpg
291
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 91.jpg
292
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 101.jpg
293
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 121.jpg
294
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 141.jpg
295
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 151.jpg
296
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 161.jpg
297
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 191.jpg
298
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 211.jpg
299
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 221.jpg
300
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 231.jpg
301
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 241.jpg
302
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 251.jpg
303
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 301.jpg
304
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 311.jpg
305
+ CAD120_Subject3_rgbd_images_microwaving_food_sequence 331.jpg
306
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 1.jpg
307
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 11.jpg
308
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 71.jpg
309
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 81.jpg
310
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 101.jpg
311
+ CAD120_Subject3_rgbd_images_picking_objects_sequence 111.jpg
312
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 1.jpg
313
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 11.jpg
314
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 41.jpg
315
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 51.jpg
316
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 61.jpg
317
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 71.jpg
318
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 81.jpg
319
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 91.jpg
320
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 101.jpg
321
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 111.jpg
322
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 161.jpg
323
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 171.jpg
324
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 181.jpg
325
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 191.jpg
326
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 201.jpg
327
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 221.jpg
328
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 231.jpg
329
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 241.jpg
330
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 251.jpg
331
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 291.jpg
332
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 321.jpg
333
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 331.jpg
334
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 371.jpg
335
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 381.jpg
336
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 391.jpg
337
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 401.jpg
338
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 411.jpg
339
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 421.jpg
340
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 431.jpg
341
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 471.jpg
342
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 481.jpg
343
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 491.jpg
344
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 501.jpg
345
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 521.jpg
346
+ CAD120_Subject3_rgbd_images_stacking_objects_sequence 551.jpg
347
+ CAD120_Subject3_rgbd_images_taking_food_sequence 21.jpg
348
+ CAD120_Subject3_rgbd_images_taking_food_sequence 31.jpg
349
+ CAD120_Subject3_rgbd_images_taking_food_sequence 51.jpg
350
+ CAD120_Subject3_rgbd_images_taking_food_sequence 81.jpg
351
+ CAD120_Subject3_rgbd_images_taking_food_sequence 101.jpg
352
+ CAD120_Subject3_rgbd_images_taking_food_sequence 111.jpg
353
+ CAD120_Subject3_rgbd_images_taking_food_sequence 141.jpg
354
+ CAD120_Subject3_rgbd_images_taking_food_sequence 161.jpg
355
+ CAD120_Subject3_rgbd_images_taking_food_sequence 181.jpg
356
+ CAD120_Subject3_rgbd_images_taking_food_sequence 201.jpg
357
+ CAD120_Subject3_rgbd_images_taking_food_sequence 211.jpg
358
+ CAD120_Subject3_rgbd_images_taking_food_sequence 231.jpg
359
+ CAD120_Subject3_rgbd_images_taking_food_sequence 241.jpg
360
+ CAD120_Subject3_rgbd_images_taking_food_sequence 251.jpg
361
+ CAD120_Subject3_rgbd_images_taking_food_sequence 271.jpg
362
+ CAD120_Subject3_rgbd_images_taking_food_sequence 291.jpg
363
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 1.jpg
364
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 11.jpg
365
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 21.jpg
366
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 31.jpg
367
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 61.jpg
368
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 71.jpg
369
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 101.jpg
370
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 111.jpg
371
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 131.jpg
372
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 151.jpg
373
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 211.jpg
374
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 231.jpg
375
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 261.jpg
376
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 301.jpg
377
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 311.jpg
378
+ CAD120_Subject3_rgbd_images_taking_medicine_sequence 321.jpg
379
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 11.jpg
380
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 21.jpg
381
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 31.jpg
382
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 41.jpg
383
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 61.jpg
384
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 71.jpg
385
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 81.jpg
386
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 101.jpg
387
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 111.jpg
388
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 131.jpg
389
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 141.jpg
390
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 151.jpg
391
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 171.jpg
392
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 191.jpg
393
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 221.jpg
394
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 231.jpg
395
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 251.jpg
396
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 261.jpg
397
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 281.jpg
398
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 301.jpg
399
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 321.jpg
400
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 341.jpg
401
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 351.jpg
402
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 451.jpg
403
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 471.jpg
404
+ CAD120_Subject3_rgbd_images_unstacking_objects_sequence 501.jpg
405
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 1.jpg
406
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 11.jpg
407
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 21.jpg
408
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 31.jpg
409
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 71.jpg
410
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 91.jpg
411
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 101.jpg
412
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 111.jpg
413
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 121.jpg
414
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 131.jpg
415
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 141.jpg
416
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 161.jpg
417
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 171.jpg
418
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 181.jpg
419
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 201.jpg
420
+ CAD120_Subject4_rgbd_images_arranging_objects_sequence 211.jpg
421
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 11.jpg
422
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 21.jpg
423
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 31.jpg
424
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 51.jpg
425
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 71.jpg
426
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 81.jpg
427
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 141.jpg
428
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 171.jpg
429
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 191.jpg
430
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 201.jpg
431
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 211.jpg
432
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 221.jpg
433
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 241.jpg
434
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 251.jpg
435
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 261.jpg
436
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 271.jpg
437
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 291.jpg
438
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 301.jpg
439
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 331.jpg
440
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 341.jpg
441
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 351.jpg
442
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 371.jpg
443
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 381.jpg
444
+ CAD120_Subject4_rgbd_images_cleaning_objects_sequence 401.jpg
445
+ CAD120_Subject4_rgbd_images_having_meal_sequence 1.jpg
446
+ CAD120_Subject4_rgbd_images_having_meal_sequence 21.jpg
447
+ CAD120_Subject4_rgbd_images_having_meal_sequence 41.jpg
448
+ CAD120_Subject4_rgbd_images_having_meal_sequence 71.jpg
449
+ CAD120_Subject4_rgbd_images_having_meal_sequence 101.jpg
450
+ CAD120_Subject4_rgbd_images_having_meal_sequence 121.jpg
451
+ CAD120_Subject4_rgbd_images_having_meal_sequence 141.jpg
452
+ CAD120_Subject4_rgbd_images_having_meal_sequence 151.jpg
453
+ CAD120_Subject4_rgbd_images_having_meal_sequence 161.jpg
454
+ CAD120_Subject4_rgbd_images_having_meal_sequence 171.jpg
455
+ CAD120_Subject4_rgbd_images_having_meal_sequence 181.jpg
456
+ CAD120_Subject4_rgbd_images_having_meal_sequence 201.jpg
457
+ CAD120_Subject4_rgbd_images_having_meal_sequence 221.jpg
458
+ CAD120_Subject4_rgbd_images_having_meal_sequence 231.jpg
459
+ CAD120_Subject4_rgbd_images_having_meal_sequence 241.jpg
460
+ CAD120_Subject4_rgbd_images_having_meal_sequence 251.jpg
461
+ CAD120_Subject4_rgbd_images_having_meal_sequence 271.jpg
462
+ CAD120_Subject4_rgbd_images_having_meal_sequence 311.jpg
463
+ CAD120_Subject4_rgbd_images_having_meal_sequence 331.jpg
464
+ CAD120_Subject4_rgbd_images_having_meal_sequence 341.jpg
465
+ CAD120_Subject4_rgbd_images_having_meal_sequence 351.jpg
466
+ CAD120_Subject4_rgbd_images_having_meal_sequence 361.jpg
467
+ CAD120_Subject4_rgbd_images_having_meal_sequence 371.jpg
468
+ CAD120_Subject4_rgbd_images_having_meal_sequence 401.jpg
469
+ CAD120_Subject4_rgbd_images_having_meal_sequence 411.jpg
470
+ CAD120_Subject4_rgbd_images_having_meal_sequence 461.jpg
471
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 1.jpg
472
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 31.jpg
473
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 61.jpg
474
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 71.jpg
475
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 81.jpg
476
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 91.jpg
477
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 111.jpg
478
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 121.jpg
479
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 151.jpg
480
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 191.jpg
481
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 211.jpg
482
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 221.jpg
483
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 231.jpg
484
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 241.jpg
485
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 251.jpg
486
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 271.jpg
487
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 281.jpg
488
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 291.jpg
489
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 301.jpg
490
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 321.jpg
491
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 341.jpg
492
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 351.jpg
493
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 371.jpg
494
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 381.jpg
495
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 391.jpg
496
+ CAD120_Subject4_rgbd_images_making_cereal_sequence 401.jpg
497
+ CAD120_Subject4_rgbd_images_microwaving_food_sequence 11.jpg
498
+ CAD120_Subject4_rgbd_images_microwaving_food_sequence 21.jpg
499
+ CAD120_Subject4_rgbd_images_microwaving_food_sequence 61.jpg
500
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 11.jpg
501
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 21.jpg
502
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 31.jpg
503
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 61.jpg
504
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 71.jpg
505
+ CAD120_Subject4_rgbd_images_picking_objects_sequence 91.jpg
506
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 1.jpg
507
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 11.jpg
508
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 21.jpg
509
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 31.jpg
510
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 61.jpg
511
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 71.jpg
512
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 111.jpg
513
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 151.jpg
514
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 161.jpg
515
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 171.jpg
516
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 181.jpg
517
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 191.jpg
518
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 211.jpg
519
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 221.jpg
520
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 231.jpg
521
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 241.jpg
522
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 251.jpg
523
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 271.jpg
524
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 291.jpg
525
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 311.jpg
526
+ CAD120_Subject4_rgbd_images_stacking_objects_sequence 321.jpg
527
+ CAD120_Subject4_rgbd_images_taking_food_sequence 1.jpg
528
+ CAD120_Subject4_rgbd_images_taking_food_sequence 11.jpg
529
+ CAD120_Subject4_rgbd_images_taking_food_sequence 21.jpg
530
+ CAD120_Subject4_rgbd_images_taking_food_sequence 41.jpg
531
+ CAD120_Subject4_rgbd_images_taking_food_sequence 71.jpg
532
+ CAD120_Subject4_rgbd_images_taking_food_sequence 91.jpg
533
+ CAD120_Subject4_rgbd_images_taking_food_sequence 101.jpg
534
+ CAD120_Subject4_rgbd_images_taking_food_sequence 141.jpg
535
+ CAD120_Subject4_rgbd_images_taking_food_sequence 151.jpg
536
+ CAD120_Subject4_rgbd_images_taking_food_sequence 161.jpg
537
+ CAD120_Subject4_rgbd_images_taking_food_sequence 171.jpg
538
+ CAD120_Subject4_rgbd_images_taking_food_sequence 181.jpg
539
+ CAD120_Subject4_rgbd_images_taking_food_sequence 201.jpg
540
+ CAD120_Subject4_rgbd_images_taking_food_sequence 241.jpg
541
+ CAD120_Subject4_rgbd_images_taking_food_sequence 261.jpg
542
+ CAD120_Subject4_rgbd_images_taking_food_sequence 281.jpg
543
+ CAD120_Subject4_rgbd_images_taking_food_sequence 311.jpg
544
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 31.jpg
545
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 51.jpg
546
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 61.jpg
547
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 71.jpg
548
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 81.jpg
549
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 91.jpg
550
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 121.jpg
551
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 141.jpg
552
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 151.jpg
553
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 171.jpg
554
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 181.jpg
555
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 191.jpg
556
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 221.jpg
557
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 231.jpg
558
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 241.jpg
559
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 261.jpg
560
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 271.jpg
561
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 281.jpg
562
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 291.jpg
563
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 301.jpg
564
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 321.jpg
565
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 331.jpg
566
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 341.jpg
567
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 371.jpg
568
+ CAD120_Subject4_rgbd_images_taking_medicine_sequence 381.jpg
569
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 1.jpg
570
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 31.jpg
571
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 51.jpg
572
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 61.jpg
573
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 71.jpg
574
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 81.jpg
575
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 91.jpg
576
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 111.jpg
577
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 131.jpg
578
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 141.jpg
579
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 151.jpg
580
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 161.jpg
581
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 171.jpg
582
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 191.jpg
583
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 201.jpg
584
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 221.jpg
585
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 241.jpg
586
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 251.jpg
587
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 261.jpg
588
+ CAD120_Subject4_rgbd_images_unstacking_objects_sequence 271.jpg
589
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 11.jpg
590
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 21.jpg
591
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 31.jpg
592
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 41.jpg
593
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 71.jpg
594
+ CAD120_Subject5_rgbd_images_arranging_objects_sequence 81.jpg
595
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 21.jpg
596
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 31.jpg
597
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 51.jpg
598
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 71.jpg
599
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 81.jpg
600
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 101.jpg
601
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 111.jpg
602
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 121.jpg
603
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 131.jpg
604
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 141.jpg
605
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 151.jpg
606
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 171.jpg
607
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 201.jpg
608
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 211.jpg
609
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 221.jpg
610
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 231.jpg
611
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 251.jpg
612
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 271.jpg
613
+ CAD120_Subject5_rgbd_images_cleaning_objects_sequence 281.jpg
614
+ CAD120_Subject5_rgbd_images_having_meal_sequence 1.jpg
615
+ CAD120_Subject5_rgbd_images_having_meal_sequence 11.jpg
616
+ CAD120_Subject5_rgbd_images_having_meal_sequence 21.jpg
617
+ CAD120_Subject5_rgbd_images_having_meal_sequence 31.jpg
618
+ CAD120_Subject5_rgbd_images_having_meal_sequence 41.jpg
619
+ CAD120_Subject5_rgbd_images_having_meal_sequence 61.jpg
620
+ CAD120_Subject5_rgbd_images_having_meal_sequence 101.jpg
621
+ CAD120_Subject5_rgbd_images_having_meal_sequence 111.jpg
622
+ CAD120_Subject5_rgbd_images_having_meal_sequence 151.jpg
623
+ CAD120_Subject5_rgbd_images_having_meal_sequence 201.jpg
624
+ CAD120_Subject5_rgbd_images_having_meal_sequence 221.jpg
625
+ CAD120_Subject5_rgbd_images_having_meal_sequence 251.jpg
626
+ CAD120_Subject5_rgbd_images_having_meal_sequence 271.jpg
627
+ CAD120_Subject5_rgbd_images_having_meal_sequence 281.jpg
628
+ CAD120_Subject5_rgbd_images_having_meal_sequence 351.jpg
629
+ CAD120_Subject5_rgbd_images_having_meal_sequence 361.jpg
630
+ CAD120_Subject5_rgbd_images_having_meal_sequence 371.jpg
631
+ CAD120_Subject5_rgbd_images_having_meal_sequence 381.jpg
632
+ CAD120_Subject5_rgbd_images_having_meal_sequence 421.jpg
633
+ CAD120_Subject5_rgbd_images_having_meal_sequence 431.jpg
634
+ CAD120_Subject5_rgbd_images_having_meal_sequence 441.jpg
635
+ CAD120_Subject5_rgbd_images_having_meal_sequence 451.jpg
636
+ CAD120_Subject5_rgbd_images_having_meal_sequence 461.jpg
637
+ CAD120_Subject5_rgbd_images_having_meal_sequence 471.jpg
638
+ CAD120_Subject5_rgbd_images_having_meal_sequence 481.jpg
639
+ CAD120_Subject5_rgbd_images_having_meal_sequence 531.jpg
640
+ CAD120_Subject5_rgbd_images_having_meal_sequence 541.jpg
641
+ CAD120_Subject5_rgbd_images_having_meal_sequence 551.jpg
642
+ CAD120_Subject5_rgbd_images_having_meal_sequence 561.jpg
643
+ CAD120_Subject5_rgbd_images_having_meal_sequence 581.jpg
644
+ CAD120_Subject5_rgbd_images_having_meal_sequence 621.jpg
645
+ CAD120_Subject5_rgbd_images_having_meal_sequence 641.jpg
646
+ CAD120_Subject5_rgbd_images_having_meal_sequence 651.jpg
647
+ CAD120_Subject5_rgbd_images_having_meal_sequence 671.jpg
648
+ CAD120_Subject5_rgbd_images_having_meal_sequence 681.jpg
649
+ CAD120_Subject5_rgbd_images_having_meal_sequence 691.jpg
650
+ CAD120_Subject5_rgbd_images_having_meal_sequence 701.jpg
651
+ CAD120_Subject5_rgbd_images_having_meal_sequence 711.jpg
652
+ CAD120_Subject5_rgbd_images_having_meal_sequence 721.jpg
653
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 11.jpg
654
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 31.jpg
655
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 41.jpg
656
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 71.jpg
657
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 91.jpg
658
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 101.jpg
659
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 121.jpg
660
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 161.jpg
661
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 171.jpg
662
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 181.jpg
663
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 191.jpg
664
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 221.jpg
665
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 241.jpg
666
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 271.jpg
667
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 281.jpg
668
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 301.jpg
669
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 331.jpg
670
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 341.jpg
671
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 351.jpg
672
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 361.jpg
673
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 371.jpg
674
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 381.jpg
675
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 391.jpg
676
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 411.jpg
677
+ CAD120_Subject5_rgbd_images_making_cereal_sequence 441.jpg
678
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 1.jpg
679
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 11.jpg
680
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 21.jpg
681
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 31.jpg
682
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 41.jpg
683
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 61.jpg
684
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 81.jpg
685
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 91.jpg
686
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 101.jpg
687
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 111.jpg
688
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 121.jpg
689
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 131.jpg
690
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 161.jpg
691
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 171.jpg
692
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 181.jpg
693
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 191.jpg
694
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 231.jpg
695
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 241.jpg
696
+ CAD120_Subject5_rgbd_images_microwaving_food_sequence 271.jpg
697
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 1.jpg
698
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 11.jpg
699
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 21.jpg
700
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 51.jpg
701
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 81.jpg
702
+ CAD120_Subject5_rgbd_images_picking_objects_sequence 121.jpg
703
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 1.jpg
704
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 11.jpg
705
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 41.jpg
706
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 51.jpg
707
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 61.jpg
708
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 81.jpg
709
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 91.jpg
710
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 101.jpg
711
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 111.jpg
712
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 141.jpg
713
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 161.jpg
714
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 191.jpg
715
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 281.jpg
716
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 291.jpg
717
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 301.jpg
718
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 311.jpg
719
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 331.jpg
720
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 351.jpg
721
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 361.jpg
722
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 371.jpg
723
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 381.jpg
724
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 391.jpg
725
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 401.jpg
726
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 421.jpg
727
+ CAD120_Subject5_rgbd_images_stacking_objects_sequence 431.jpg
728
+ CAD120_Subject5_rgbd_images_taking_food_sequence 11.jpg
729
+ CAD120_Subject5_rgbd_images_taking_food_sequence 51.jpg
730
+ CAD120_Subject5_rgbd_images_taking_food_sequence 71.jpg
731
+ CAD120_Subject5_rgbd_images_taking_food_sequence 81.jpg
732
+ CAD120_Subject5_rgbd_images_taking_food_sequence 101.jpg
733
+ CAD120_Subject5_rgbd_images_taking_food_sequence 111.jpg
734
+ CAD120_Subject5_rgbd_images_taking_food_sequence 121.jpg
735
+ CAD120_Subject5_rgbd_images_taking_food_sequence 131.jpg
736
+ CAD120_Subject5_rgbd_images_taking_food_sequence 141.jpg
737
+ CAD120_Subject5_rgbd_images_taking_food_sequence 171.jpg
738
+ CAD120_Subject5_rgbd_images_taking_food_sequence 191.jpg
739
+ CAD120_Subject5_rgbd_images_taking_food_sequence 211.jpg
740
+ CAD120_Subject5_rgbd_images_taking_food_sequence 221.jpg
741
+ CAD120_Subject5_rgbd_images_taking_food_sequence 231.jpg
742
+ CAD120_Subject5_rgbd_images_taking_food_sequence 241.jpg
743
+ CAD120_Subject5_rgbd_images_taking_food_sequence 261.jpg
744
+ CAD120_Subject5_rgbd_images_taking_food_sequence 311.jpg
745
+ CAD120_Subject5_rgbd_images_taking_food_sequence 321.jpg
746
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 1.jpg
747
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 41.jpg
748
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 51.jpg
749
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 61.jpg
750
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 81.jpg
751
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 91.jpg
752
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 101.jpg
753
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 111.jpg
754
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 121.jpg
755
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 131.jpg
756
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 151.jpg
757
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 161.jpg
758
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 171.jpg
759
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 181.jpg
760
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 211.jpg
761
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 231.jpg
762
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 241.jpg
763
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 251.jpg
764
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 271.jpg
765
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 281.jpg
766
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 311.jpg
767
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 331.jpg
768
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 341.jpg
769
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 351.jpg
770
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 361.jpg
771
+ CAD120_Subject5_rgbd_images_taking_medicine_sequence 391.jpg
772
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 11.jpg
773
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 51.jpg
774
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 61.jpg
775
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 71.jpg
776
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 151.jpg
777
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 161.jpg
778
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 171.jpg
779
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 181.jpg
780
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 191.jpg
781
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 201.jpg
782
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 211.jpg
783
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 241.jpg
784
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 261.jpg
785
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 271.jpg
786
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 281.jpg
787
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 321.jpg
788
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 331.jpg
789
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 371.jpg
790
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 381.jpg
791
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 401.jpg
792
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 411.jpg
793
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 421.jpg
794
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 431.jpg
795
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 441.jpg
796
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 461.jpg
797
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 481.jpg
798
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 491.jpg
799
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 511.jpg
800
+ CAD120_Subject5_rgbd_images_unstacking_objects_sequence 521.jpg
external/WildCamera/splits/cityscapes_test.txt ADDED
@@ -0,0 +1,800 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ berlin berlin_000000_000019
2
+ berlin berlin_000001_000019
3
+ berlin berlin_000002_000019
4
+ berlin berlin_000004_000019
5
+ berlin berlin_000005_000019
6
+ berlin berlin_000009_000019
7
+ berlin berlin_000013_000019
8
+ berlin berlin_000014_000019
9
+ berlin berlin_000016_000019
10
+ berlin berlin_000018_000019
11
+ berlin berlin_000019_000019
12
+ berlin berlin_000020_000019
13
+ berlin berlin_000021_000019
14
+ berlin berlin_000022_000019
15
+ berlin berlin_000023_000019
16
+ berlin berlin_000026_000019
17
+ berlin berlin_000027_000019
18
+ berlin berlin_000029_000019
19
+ berlin berlin_000033_000019
20
+ berlin berlin_000036_000019
21
+ berlin berlin_000037_000019
22
+ berlin berlin_000038_000019
23
+ berlin berlin_000039_000019
24
+ berlin berlin_000040_000019
25
+ berlin berlin_000041_000019
26
+ berlin berlin_000042_000019
27
+ berlin berlin_000044_000019
28
+ berlin berlin_000046_000019
29
+ berlin berlin_000051_000019
30
+ berlin berlin_000052_000019
31
+ berlin berlin_000053_000019
32
+ berlin berlin_000056_000019
33
+ berlin berlin_000058_000019
34
+ berlin berlin_000063_000019
35
+ berlin berlin_000066_000019
36
+ berlin berlin_000068_000019
37
+ berlin berlin_000069_000019
38
+ berlin berlin_000072_000019
39
+ berlin berlin_000073_000019
40
+ berlin berlin_000074_000019
41
+ berlin berlin_000079_000019
42
+ berlin berlin_000080_000019
43
+ berlin berlin_000081_000019
44
+ berlin berlin_000082_000019
45
+ berlin berlin_000083_000019
46
+ berlin berlin_000085_000019
47
+ berlin berlin_000086_000019
48
+ berlin berlin_000087_000019
49
+ berlin berlin_000088_000019
50
+ berlin berlin_000089_000019
51
+ berlin berlin_000090_000019
52
+ berlin berlin_000091_000019
53
+ berlin berlin_000092_000019
54
+ berlin berlin_000093_000019
55
+ berlin berlin_000094_000019
56
+ berlin berlin_000096_000019
57
+ berlin berlin_000097_000019
58
+ berlin berlin_000098_000019
59
+ berlin berlin_000099_000019
60
+ berlin berlin_000102_000019
61
+ berlin berlin_000106_000019
62
+ berlin berlin_000108_000019
63
+ berlin berlin_000109_000019
64
+ berlin berlin_000114_000019
65
+ berlin berlin_000116_000019
66
+ berlin berlin_000117_000019
67
+ berlin berlin_000122_000019
68
+ berlin berlin_000123_000019
69
+ berlin berlin_000126_000019
70
+ berlin berlin_000128_000019
71
+ berlin berlin_000131_000019
72
+ berlin berlin_000135_000019
73
+ berlin berlin_000137_000019
74
+ berlin berlin_000138_000019
75
+ berlin berlin_000141_000019
76
+ berlin berlin_000144_000019
77
+ berlin berlin_000145_000019
78
+ berlin berlin_000146_000019
79
+ berlin berlin_000147_000019
80
+ berlin berlin_000148_000019
81
+ berlin berlin_000149_000019
82
+ berlin berlin_000153_000019
83
+ berlin berlin_000155_000019
84
+ berlin berlin_000156_000019
85
+ berlin berlin_000157_000019
86
+ berlin berlin_000158_000019
87
+ berlin berlin_000162_000019
88
+ berlin berlin_000163_000019
89
+ berlin berlin_000164_000019
90
+ berlin berlin_000169_000019
91
+ berlin berlin_000171_000019
92
+ berlin berlin_000172_000019
93
+ berlin berlin_000178_000019
94
+ berlin berlin_000181_000019
95
+ berlin berlin_000182_000019
96
+ berlin berlin_000184_000019
97
+ berlin berlin_000185_000019
98
+ berlin berlin_000186_000019
99
+ berlin berlin_000187_000019
100
+ berlin berlin_000188_000019
101
+ berlin berlin_000192_000019
102
+ berlin berlin_000195_000019
103
+ berlin berlin_000196_000019
104
+ berlin berlin_000198_000019
105
+ berlin berlin_000203_000019
106
+ berlin berlin_000206_000019
107
+ berlin berlin_000211_000019
108
+ berlin berlin_000212_000019
109
+ berlin berlin_000215_000019
110
+ berlin berlin_000217_000019
111
+ berlin berlin_000222_000019
112
+ berlin berlin_000224_000019
113
+ berlin berlin_000226_000019
114
+ berlin berlin_000227_000019
115
+ berlin berlin_000229_000019
116
+ berlin berlin_000230_000019
117
+ berlin berlin_000233_000019
118
+ berlin berlin_000235_000019
119
+ berlin berlin_000236_000019
120
+ berlin berlin_000238_000019
121
+ berlin berlin_000239_000019
122
+ berlin berlin_000241_000019
123
+ berlin berlin_000244_000019
124
+ berlin berlin_000247_000019
125
+ berlin berlin_000248_000019
126
+ berlin berlin_000249_000019
127
+ berlin berlin_000255_000019
128
+ berlin berlin_000258_000019
129
+ berlin berlin_000262_000019
130
+ berlin berlin_000270_000019
131
+ berlin berlin_000271_000019
132
+ berlin berlin_000272_000019
133
+ berlin berlin_000275_000019
134
+ berlin berlin_000277_000019
135
+ berlin berlin_000278_000019
136
+ berlin berlin_000279_000019
137
+ berlin berlin_000280_000019
138
+ berlin berlin_000283_000019
139
+ berlin berlin_000286_000019
140
+ berlin berlin_000287_000019
141
+ berlin berlin_000292_000019
142
+ berlin berlin_000294_000019
143
+ berlin berlin_000299_000019
144
+ berlin berlin_000302_000019
145
+ berlin berlin_000305_000019
146
+ berlin berlin_000306_000019
147
+ berlin berlin_000311_000019
148
+ berlin berlin_000312_000019
149
+ berlin berlin_000318_000019
150
+ berlin berlin_000319_000019
151
+ berlin berlin_000326_000019
152
+ berlin berlin_000327_000019
153
+ berlin berlin_000328_000019
154
+ berlin berlin_000329_000019
155
+ berlin berlin_000331_000019
156
+ berlin berlin_000332_000019
157
+ berlin berlin_000333_000019
158
+ berlin berlin_000335_000019
159
+ berlin berlin_000338_000019
160
+ berlin berlin_000339_000019
161
+ berlin berlin_000340_000019
162
+ berlin berlin_000341_000019
163
+ berlin berlin_000343_000019
164
+ berlin berlin_000344_000019
165
+ berlin berlin_000346_000019
166
+ berlin berlin_000347_000019
167
+ berlin berlin_000348_000019
168
+ berlin berlin_000349_000019
169
+ berlin berlin_000350_000019
170
+ berlin berlin_000353_000019
171
+ berlin berlin_000354_000019
172
+ berlin berlin_000355_000019
173
+ berlin berlin_000356_000019
174
+ berlin berlin_000361_000019
175
+ berlin berlin_000363_000019
176
+ berlin berlin_000365_000019
177
+ berlin berlin_000366_000019
178
+ berlin berlin_000368_000019
179
+ berlin berlin_000371_000019
180
+ berlin berlin_000372_000019
181
+ berlin berlin_000374_000019
182
+ berlin berlin_000375_000019
183
+ berlin berlin_000376_000019
184
+ berlin berlin_000377_000019
185
+ berlin berlin_000380_000019
186
+ berlin berlin_000381_000019
187
+ berlin berlin_000382_000019
188
+ berlin berlin_000385_000019
189
+ berlin berlin_000387_000019
190
+ berlin berlin_000389_000019
191
+ berlin berlin_000390_000019
192
+ berlin berlin_000396_000019
193
+ berlin berlin_000397_000019
194
+ berlin berlin_000400_000019
195
+ berlin berlin_000401_000019
196
+ berlin berlin_000403_000019
197
+ berlin berlin_000404_000019
198
+ berlin berlin_000407_000019
199
+ berlin berlin_000409_000019
200
+ berlin berlin_000410_000019
201
+ berlin berlin_000412_000019
202
+ berlin berlin_000415_000019
203
+ berlin berlin_000416_000019
204
+ berlin berlin_000417_000019
205
+ berlin berlin_000418_000019
206
+ berlin berlin_000421_000019
207
+ berlin berlin_000422_000019
208
+ berlin berlin_000424_000019
209
+ berlin berlin_000425_000019
210
+ berlin berlin_000427_000019
211
+ berlin berlin_000429_000019
212
+ berlin berlin_000430_000019
213
+ berlin berlin_000431_000019
214
+ berlin berlin_000432_000019
215
+ berlin berlin_000435_000019
216
+ berlin berlin_000436_000019
217
+ berlin berlin_000437_000019
218
+ berlin berlin_000440_000019
219
+ berlin berlin_000444_000019
220
+ berlin berlin_000445_000019
221
+ berlin berlin_000447_000019
222
+ berlin berlin_000451_000019
223
+ berlin berlin_000453_000019
224
+ berlin berlin_000454_000019
225
+ berlin berlin_000456_000019
226
+ berlin berlin_000458_000019
227
+ berlin berlin_000461_000019
228
+ berlin berlin_000465_000019
229
+ berlin berlin_000466_000019
230
+ berlin berlin_000468_000019
231
+ berlin berlin_000469_000019
232
+ berlin berlin_000471_000019
233
+ berlin berlin_000472_000019
234
+ berlin berlin_000473_000019
235
+ berlin berlin_000479_000019
236
+ berlin berlin_000482_000019
237
+ berlin berlin_000484_000019
238
+ berlin berlin_000485_000019
239
+ berlin berlin_000486_000019
240
+ berlin berlin_000490_000019
241
+ berlin berlin_000491_000019
242
+ berlin berlin_000492_000019
243
+ berlin berlin_000493_000019
244
+ berlin berlin_000498_000019
245
+ berlin berlin_000502_000019
246
+ berlin berlin_000503_000019
247
+ berlin berlin_000504_000019
248
+ berlin berlin_000505_000019
249
+ berlin berlin_000507_000019
250
+ berlin berlin_000510_000019
251
+ berlin berlin_000511_000019
252
+ berlin berlin_000512_000019
253
+ berlin berlin_000514_000019
254
+ berlin berlin_000515_000019
255
+ berlin berlin_000516_000019
256
+ berlin berlin_000518_000019
257
+ berlin berlin_000519_000019
258
+ berlin berlin_000521_000019
259
+ berlin berlin_000522_000019
260
+ berlin berlin_000523_000019
261
+ berlin berlin_000524_000019
262
+ berlin berlin_000525_000019
263
+ berlin berlin_000527_000019
264
+ berlin berlin_000529_000019
265
+ berlin berlin_000530_000019
266
+ berlin berlin_000531_000019
267
+ berlin berlin_000532_000019
268
+ berlin berlin_000536_000019
269
+ berlin berlin_000539_000019
270
+ berlin berlin_000541_000019
271
+ bielefeld bielefeld_000000_001187
272
+ bielefeld bielefeld_000000_001505
273
+ bielefeld bielefeld_000000_002308
274
+ bielefeld bielefeld_000000_003406
275
+ bielefeld bielefeld_000000_003546
276
+ bielefeld bielefeld_000000_003731
277
+ bielefeld bielefeld_000000_005068
278
+ bielefeld bielefeld_000000_005584
279
+ bielefeld bielefeld_000000_006239
280
+ bielefeld bielefeld_000000_006802
281
+ bielefeld bielefeld_000000_007030
282
+ bielefeld bielefeld_000000_008279
283
+ bielefeld bielefeld_000000_008581
284
+ bielefeld bielefeld_000000_008800
285
+ bielefeld bielefeld_000000_009728
286
+ bielefeld bielefeld_000000_009928
287
+ bielefeld bielefeld_000000_011367
288
+ bielefeld bielefeld_000000_011831
289
+ bielefeld bielefeld_000000_012080
290
+ bielefeld bielefeld_000000_012788
291
+ bielefeld bielefeld_000000_013570
292
+ bielefeld bielefeld_000000_013814
293
+ bielefeld bielefeld_000000_015301
294
+ bielefeld bielefeld_000000_015867
295
+ bielefeld bielefeld_000000_016019
296
+ bielefeld bielefeld_000000_017051
297
+ bielefeld bielefeld_000000_017774
298
+ bielefeld bielefeld_000000_018102
299
+ bielefeld bielefeld_000000_018644
300
+ bielefeld bielefeld_000000_019416
301
+ bielefeld bielefeld_000000_020757
302
+ bielefeld bielefeld_000000_020900
303
+ bielefeld bielefeld_000000_021341
304
+ bielefeld bielefeld_000000_021381
305
+ bielefeld bielefeld_000000_022261
306
+ bielefeld bielefeld_000000_022835
307
+ bielefeld bielefeld_000000_025061
308
+ bielefeld bielefeld_000000_025748
309
+ bielefeld bielefeld_000000_026053
310
+ bielefeld bielefeld_000000_026550
311
+ bielefeld bielefeld_000000_026660
312
+ bielefeld bielefeld_000000_026823
313
+ bielefeld bielefeld_000000_027221
314
+ bielefeld bielefeld_000000_027586
315
+ bielefeld bielefeld_000000_027928
316
+ bielefeld bielefeld_000000_028148
317
+ bielefeld bielefeld_000000_028550
318
+ bielefeld bielefeld_000000_030038
319
+ bielefeld bielefeld_000000_030958
320
+ bielefeld bielefeld_000000_031244
321
+ bielefeld bielefeld_000000_032388
322
+ bielefeld bielefeld_000000_033770
323
+ bielefeld bielefeld_000000_035537
324
+ bielefeld bielefeld_000000_035879
325
+ bielefeld bielefeld_000000_036362
326
+ bielefeld bielefeld_000000_036732
327
+ bielefeld bielefeld_000000_037016
328
+ bielefeld bielefeld_000000_037422
329
+ bielefeld bielefeld_000000_038924
330
+ bielefeld bielefeld_000000_039221
331
+ bielefeld bielefeld_000000_040035
332
+ bielefeld bielefeld_000000_040472
333
+ bielefeld bielefeld_000000_041014
334
+ bielefeld bielefeld_000000_041142
335
+ bielefeld bielefeld_000000_041444
336
+ bielefeld bielefeld_000000_042403
337
+ bielefeld bielefeld_000000_042571
338
+ bielefeld bielefeld_000000_042717
339
+ bielefeld bielefeld_000000_043100
340
+ bielefeld bielefeld_000000_043389
341
+ bielefeld bielefeld_000000_043610
342
+ bielefeld bielefeld_000000_044085
343
+ bielefeld bielefeld_000000_045117
344
+ bielefeld bielefeld_000000_045232
345
+ bielefeld bielefeld_000000_046023
346
+ bielefeld bielefeld_000000_046212
347
+ bielefeld bielefeld_000000_046495
348
+ bielefeld bielefeld_000000_047918
349
+ bielefeld bielefeld_000000_048227
350
+ bielefeld bielefeld_000000_048518
351
+ bielefeld bielefeld_000000_049446
352
+ bielefeld bielefeld_000000_050586
353
+ bielefeld bielefeld_000000_051894
354
+ bielefeld bielefeld_000000_052155
355
+ bielefeld bielefeld_000000_053384
356
+ bielefeld bielefeld_000000_053583
357
+ bielefeld bielefeld_000000_055145
358
+ bielefeld bielefeld_000000_056175
359
+ bielefeld bielefeld_000000_056226
360
+ bielefeld bielefeld_000000_056603
361
+ bielefeld bielefeld_000000_056866
362
+ bielefeld bielefeld_000000_058374
363
+ bielefeld bielefeld_000000_058934
364
+ bielefeld bielefeld_000000_059303
365
+ bielefeld bielefeld_000000_059766
366
+ bielefeld bielefeld_000000_059842
367
+ bielefeld bielefeld_000000_060786
368
+ bielefeld bielefeld_000000_061094
369
+ bielefeld bielefeld_000000_062121
370
+ bielefeld bielefeld_000000_063623
371
+ bielefeld bielefeld_000000_063939
372
+ bielefeld bielefeld_000000_064271
373
+ bielefeld bielefeld_000000_064805
374
+ bielefeld bielefeld_000000_064910
375
+ bielefeld bielefeld_000000_066195
376
+ bielefeld bielefeld_000000_066405
377
+ bielefeld bielefeld_000000_066495
378
+ bonn bonn_000001_000019
379
+ bonn bonn_000002_000019
380
+ bonn bonn_000005_000019
381
+ bonn bonn_000006_000019
382
+ bonn bonn_000010_000019
383
+ bonn bonn_000013_000019
384
+ bonn bonn_000016_000019
385
+ bonn bonn_000020_000019
386
+ bonn bonn_000021_000019
387
+ bonn bonn_000022_000019
388
+ bonn bonn_000026_000019
389
+ bonn bonn_000027_000019
390
+ bonn bonn_000028_000019
391
+ bonn bonn_000029_000019
392
+ bonn bonn_000031_000019
393
+ bonn bonn_000033_000019
394
+ bonn bonn_000034_000019
395
+ bonn bonn_000037_000019
396
+ bonn bonn_000038_000019
397
+ bonn bonn_000039_000019
398
+ bonn bonn_000040_000019
399
+ bonn bonn_000041_000019
400
+ bonn bonn_000042_000019
401
+ bonn bonn_000043_000019
402
+ bonn bonn_000044_000019
403
+ leverkusen leverkusen_000003_000019
404
+ leverkusen leverkusen_000005_000019
405
+ leverkusen leverkusen_000006_000019
406
+ leverkusen leverkusen_000009_000019
407
+ leverkusen leverkusen_000010_000019
408
+ leverkusen leverkusen_000011_000019
409
+ leverkusen leverkusen_000012_000019
410
+ leverkusen leverkusen_000017_000019
411
+ leverkusen leverkusen_000018_000019
412
+ leverkusen leverkusen_000020_000019
413
+ leverkusen leverkusen_000021_000019
414
+ leverkusen leverkusen_000022_000019
415
+ leverkusen leverkusen_000023_000019
416
+ leverkusen leverkusen_000027_000019
417
+ leverkusen leverkusen_000029_000019
418
+ leverkusen leverkusen_000030_000019
419
+ leverkusen leverkusen_000031_000019
420
+ leverkusen leverkusen_000032_000019
421
+ leverkusen leverkusen_000034_000019
422
+ leverkusen leverkusen_000035_000019
423
+ leverkusen leverkusen_000036_000019
424
+ leverkusen leverkusen_000040_000019
425
+ leverkusen leverkusen_000041_000019
426
+ leverkusen leverkusen_000042_000019
427
+ leverkusen leverkusen_000044_000019
428
+ leverkusen leverkusen_000045_000019
429
+ leverkusen leverkusen_000048_000019
430
+ leverkusen leverkusen_000049_000019
431
+ leverkusen leverkusen_000050_000019
432
+ leverkusen leverkusen_000051_000019
433
+ leverkusen leverkusen_000052_000019
434
+ leverkusen leverkusen_000054_000019
435
+ leverkusen leverkusen_000056_000019
436
+ mainz mainz_000000_000093
437
+ mainz mainz_000000_000293
438
+ mainz mainz_000000_001003
439
+ mainz mainz_000000_001068
440
+ mainz mainz_000000_001265
441
+ mainz mainz_000000_001410
442
+ mainz mainz_000000_001857
443
+ mainz mainz_000000_003250
444
+ mainz mainz_000000_007415
445
+ mainz mainz_000000_008165
446
+ mainz mainz_000000_008645
447
+ mainz mainz_000000_008871
448
+ mainz mainz_000000_009985
449
+ mainz mainz_000000_010550
450
+ mainz mainz_000000_011339
451
+ mainz mainz_000000_011879
452
+ mainz mainz_000000_011965
453
+ mainz mainz_000000_012392
454
+ mainz mainz_000000_013095
455
+ mainz mainz_000000_013437
456
+ mainz mainz_000000_013960
457
+ mainz mainz_000000_014193
458
+ mainz mainz_000000_014742
459
+ mainz mainz_000000_015052
460
+ mainz mainz_000000_015170
461
+ mainz mainz_000000_015760
462
+ mainz mainz_000000_017927
463
+ mainz mainz_000000_018249
464
+ mainz mainz_000000_018883
465
+ mainz mainz_000000_019043
466
+ mainz mainz_000000_019227
467
+ mainz mainz_000000_019847
468
+ mainz mainz_000000_020415
469
+ mainz mainz_000000_021457
470
+ mainz mainz_000000_021524
471
+ mainz mainz_000000_021735
472
+ mainz mainz_000000_022417
473
+ mainz mainz_000001_000120
474
+ mainz mainz_000001_000428
475
+ mainz mainz_000001_001509
476
+ mainz mainz_000001_002884
477
+ mainz mainz_000001_003012
478
+ mainz mainz_000001_003624
479
+ mainz mainz_000001_004132
480
+ mainz mainz_000001_004219
481
+ mainz mainz_000001_004823
482
+ mainz mainz_000001_005016
483
+ mainz mainz_000001_005163
484
+ mainz mainz_000001_005665
485
+ mainz mainz_000001_005815
486
+ mainz mainz_000001_005911
487
+ mainz mainz_000001_006194
488
+ mainz mainz_000001_007595
489
+ mainz mainz_000001_007956
490
+ mainz mainz_000001_008056
491
+ mainz mainz_000001_008264
492
+ mainz mainz_000001_008540
493
+ mainz mainz_000001_008638
494
+ mainz mainz_000001_009328
495
+ mainz mainz_000001_009867
496
+ mainz mainz_000001_009996
497
+ mainz mainz_000001_010853
498
+ mainz mainz_000001_011736
499
+ mainz mainz_000001_012541
500
+ mainz mainz_000001_014073
501
+ mainz mainz_000001_014469
502
+ mainz mainz_000001_015117
503
+ mainz mainz_000001_015508
504
+ mainz mainz_000001_015724
505
+ mainz mainz_000001_018145
506
+ mainz mainz_000001_019061
507
+ mainz mainz_000001_019593
508
+ mainz mainz_000001_020068
509
+ mainz mainz_000001_020484
510
+ mainz mainz_000001_020829
511
+ mainz mainz_000001_022125
512
+ mainz mainz_000001_022630
513
+ mainz mainz_000001_023604
514
+ mainz mainz_000001_024489
515
+ mainz mainz_000001_024718
516
+ mainz mainz_000001_025623
517
+ mainz mainz_000001_026837
518
+ mainz mainz_000001_027053
519
+ mainz mainz_000001_027377
520
+ mainz mainz_000001_027751
521
+ mainz mainz_000001_028326
522
+ mainz mainz_000001_028566
523
+ mainz mainz_000001_029755
524
+ mainz mainz_000001_029950
525
+ mainz mainz_000001_030417
526
+ mainz mainz_000001_031697
527
+ mainz mainz_000001_031946
528
+ mainz mainz_000001_032294
529
+ mainz mainz_000001_032401
530
+ mainz mainz_000001_032691
531
+ mainz mainz_000001_032767
532
+ mainz mainz_000001_032911
533
+ mainz mainz_000001_033096
534
+ mainz mainz_000001_033603
535
+ mainz mainz_000001_034209
536
+ mainz mainz_000001_034681
537
+ mainz mainz_000001_035293
538
+ mainz mainz_000001_035963
539
+ mainz mainz_000001_036240
540
+ mainz mainz_000001_036412
541
+ mainz mainz_000001_037905
542
+ mainz mainz_000001_038026
543
+ mainz mainz_000001_038191
544
+ mainz mainz_000001_038768
545
+ mainz mainz_000001_039075
546
+ mainz mainz_000001_039470
547
+ mainz mainz_000001_040367
548
+ mainz mainz_000001_041172
549
+ mainz mainz_000001_041647
550
+ mainz mainz_000001_041887
551
+ mainz mainz_000001_042121
552
+ mainz mainz_000001_042851
553
+ mainz mainz_000001_043656
554
+ mainz mainz_000001_044366
555
+ mainz mainz_000001_045385
556
+ mainz mainz_000001_045651
557
+ mainz mainz_000001_045782
558
+ mainz mainz_000001_046381
559
+ mainz mainz_000001_047546
560
+ mainz mainz_000001_047611
561
+ mainz mainz_000001_047888
562
+ mainz mainz_000001_048725
563
+ mainz mainz_000002_000061
564
+ mainz mainz_000002_000912
565
+ mainz mainz_000003_001465
566
+ mainz mainz_000003_001899
567
+ mainz mainz_000003_003558
568
+ mainz mainz_000003_003711
569
+ mainz mainz_000003_004144
570
+ mainz mainz_000003_004228
571
+ mainz mainz_000003_004774
572
+ mainz mainz_000003_005088
573
+ mainz mainz_000003_006863
574
+ mainz mainz_000003_007144
575
+ mainz mainz_000003_008690
576
+ mainz mainz_000003_008876
577
+ mainz mainz_000003_009819
578
+ mainz mainz_000003_010772
579
+ mainz mainz_000003_010924
580
+ mainz mainz_000003_011949
581
+ mainz mainz_000003_012341
582
+ mainz mainz_000003_012995
583
+ mainz mainz_000003_014083
584
+ mainz mainz_000003_014457
585
+ mainz mainz_000003_014537
586
+ mainz mainz_000003_014959
587
+ mainz mainz_000003_015411
588
+ mainz mainz_000003_015649
589
+ mainz mainz_000003_015917
590
+ mainz mainz_000003_016360
591
+ mainz mainz_000003_016542
592
+ mainz mainz_000003_017171
593
+ munich munich_000000_000019
594
+ munich munich_000002_000019
595
+ munich munich_000003_000019
596
+ munich munich_000005_000019
597
+ munich munich_000006_000019
598
+ munich munich_000009_000019
599
+ munich munich_000011_000019
600
+ munich munich_000012_000019
601
+ munich munich_000016_000019
602
+ munich munich_000017_000019
603
+ munich munich_000019_000019
604
+ munich munich_000025_000019
605
+ munich munich_000026_000019
606
+ munich munich_000028_000019
607
+ munich munich_000030_000019
608
+ munich munich_000032_000019
609
+ munich munich_000038_000019
610
+ munich munich_000040_000019
611
+ munich munich_000042_000019
612
+ munich munich_000046_000019
613
+ munich munich_000047_000019
614
+ munich munich_000050_000019
615
+ munich munich_000051_000019
616
+ munich munich_000052_000019
617
+ munich munich_000054_000019
618
+ munich munich_000055_000019
619
+ munich munich_000056_000019
620
+ munich munich_000057_000019
621
+ munich munich_000058_000019
622
+ munich munich_000062_000019
623
+ munich munich_000063_000019
624
+ munich munich_000065_000019
625
+ munich munich_000066_000019
626
+ munich munich_000067_000019
627
+ munich munich_000068_000019
628
+ munich munich_000070_000019
629
+ munich munich_000072_000019
630
+ munich munich_000075_000019
631
+ munich munich_000077_000019
632
+ munich munich_000079_000019
633
+ munich munich_000081_000019
634
+ munich munich_000082_000019
635
+ munich munich_000083_000019
636
+ munich munich_000084_000019
637
+ munich munich_000085_000019
638
+ munich munich_000088_000019
639
+ munich munich_000090_000019
640
+ munich munich_000091_000019
641
+ munich munich_000093_000019
642
+ munich munich_000096_000019
643
+ munich munich_000097_000019
644
+ munich munich_000098_000019
645
+ munich munich_000100_000019
646
+ munich munich_000101_000019
647
+ munich munich_000103_000019
648
+ munich munich_000109_000019
649
+ munich munich_000110_000019
650
+ munich munich_000111_000019
651
+ munich munich_000113_000019
652
+ munich munich_000116_000019
653
+ munich munich_000118_000019
654
+ munich munich_000120_000019
655
+ munich munich_000121_000019
656
+ munich munich_000125_000019
657
+ munich munich_000126_000019
658
+ munich munich_000128_000019
659
+ munich munich_000131_000019
660
+ munich munich_000132_000019
661
+ munich munich_000134_000019
662
+ munich munich_000135_000019
663
+ munich munich_000137_000019
664
+ munich munich_000138_000019
665
+ munich munich_000139_000019
666
+ munich munich_000140_000019
667
+ munich munich_000141_000019
668
+ munich munich_000142_000019
669
+ munich munich_000143_000019
670
+ munich munich_000144_000019
671
+ munich munich_000145_000019
672
+ munich munich_000150_000019
673
+ munich munich_000154_000019
674
+ munich munich_000155_000019
675
+ munich munich_000156_000019
676
+ munich munich_000157_000019
677
+ munich munich_000161_000019
678
+ munich munich_000164_000019
679
+ munich munich_000166_000019
680
+ munich munich_000168_000019
681
+ munich munich_000170_000019
682
+ munich munich_000173_000019
683
+ munich munich_000174_000019
684
+ munich munich_000175_000019
685
+ munich munich_000176_000019
686
+ munich munich_000178_000019
687
+ munich munich_000179_000019
688
+ munich munich_000180_000019
689
+ munich munich_000183_000019
690
+ munich munich_000186_000019
691
+ munich munich_000187_000019
692
+ munich munich_000189_000019
693
+ munich munich_000193_000019
694
+ munich munich_000195_000019
695
+ munich munich_000197_000019
696
+ munich munich_000199_000019
697
+ munich munich_000200_000019
698
+ munich munich_000203_000019
699
+ munich munich_000204_000019
700
+ munich munich_000206_000019
701
+ munich munich_000207_000019
702
+ munich munich_000209_000019
703
+ munich munich_000210_000019
704
+ munich munich_000212_000019
705
+ munich munich_000215_000019
706
+ munich munich_000216_000019
707
+ munich munich_000217_000019
708
+ munich munich_000219_000019
709
+ munich munich_000220_000019
710
+ munich munich_000221_000019
711
+ munich munich_000223_000019
712
+ munich munich_000224_000019
713
+ munich munich_000225_000019
714
+ munich munich_000227_000019
715
+ munich munich_000230_000019
716
+ munich munich_000231_000019
717
+ munich munich_000234_000019
718
+ munich munich_000235_000019
719
+ munich munich_000237_000019
720
+ munich munich_000238_000019
721
+ munich munich_000241_000019
722
+ munich munich_000245_000019
723
+ munich munich_000246_000019
724
+ munich munich_000247_000019
725
+ munich munich_000251_000019
726
+ munich munich_000255_000019
727
+ munich munich_000256_000019
728
+ munich munich_000257_000019
729
+ munich munich_000258_000019
730
+ munich munich_000260_000019
731
+ munich munich_000261_000019
732
+ munich munich_000264_000019
733
+ munich munich_000265_000019
734
+ munich munich_000266_000019
735
+ munich munich_000270_000019
736
+ munich munich_000271_000019
737
+ munich munich_000272_000019
738
+ munich munich_000273_000019
739
+ munich munich_000275_000019
740
+ munich munich_000276_000019
741
+ munich munich_000280_000019
742
+ munich munich_000281_000019
743
+ munich munich_000282_000019
744
+ munich munich_000283_000019
745
+ munich munich_000285_000019
746
+ munich munich_000286_000019
747
+ munich munich_000287_000019
748
+ munich munich_000288_000019
749
+ munich munich_000289_000019
750
+ munich munich_000290_000019
751
+ munich munich_000295_000019
752
+ munich munich_000297_000019
753
+ munich munich_000298_000019
754
+ munich munich_000300_000019
755
+ munich munich_000303_000019
756
+ munich munich_000305_000019
757
+ munich munich_000307_000019
758
+ munich munich_000309_000019
759
+ munich munich_000312_000019
760
+ munich munich_000314_000019
761
+ munich munich_000318_000019
762
+ munich munich_000320_000019
763
+ munich munich_000322_000019
764
+ munich munich_000326_000019
765
+ munich munich_000328_000019
766
+ munich munich_000329_000019
767
+ munich munich_000331_000019
768
+ munich munich_000334_000019
769
+ munich munich_000335_000019
770
+ munich munich_000336_000019
771
+ munich munich_000338_000019
772
+ munich munich_000339_000019
773
+ munich munich_000343_000019
774
+ munich munich_000347_000019
775
+ munich munich_000348_000019
776
+ munich munich_000350_000019
777
+ munich munich_000351_000019
778
+ munich munich_000352_000019
779
+ munich munich_000354_000019
780
+ munich munich_000355_000019
781
+ munich munich_000356_000019
782
+ munich munich_000357_000019
783
+ munich munich_000358_000019
784
+ munich munich_000359_000019
785
+ munich munich_000361_000019
786
+ munich munich_000362_000019
787
+ munich munich_000366_000019
788
+ munich munich_000370_000019
789
+ munich munich_000373_000019
790
+ munich munich_000375_000019
791
+ munich munich_000386_000019
792
+ munich munich_000387_000019
793
+ munich munich_000388_000019
794
+ munich munich_000389_000019
795
+ munich munich_000391_000019
796
+ munich munich_000392_000019
797
+ munich munich_000393_000019
798
+ munich munich_000394_000019
799
+ munich munich_000395_000019
800
+ munich munich_000397_000019
external/WildCamera/splits/cityscapes_train.txt ADDED
The diff for this file is too large to render. See raw diff