compvis / test /geometry /transform /test_crop3d.py
Dexter's picture
Upload folder using huggingface_hub
36c95ba verified
import pytest
import torch
from torch.autograd import gradcheck
import kornia
import kornia.testing as utils # test utils
from kornia.testing import assert_close
class TestCropAndResize3D:
def test_crop(self, device, dtype):
inp = torch.arange(0.0, 64.0, device=device, dtype=dtype).view(1, 1, 4, 4, 4)
depth, height, width = 2, 2, 2
expected = torch.tensor(
[[[[[25.1667, 27.1667], [30.5000, 32.5000]], [[46.5000, 48.5000], [51.8333, 53.8333]]]]],
device=device,
dtype=dtype,
)
boxes = torch.tensor(
[[[0, 0, 1], [3, 0, 1], [3, 2, 1], [0, 2, 1], [0, 0, 3], [3, 0, 3], [3, 2, 3], [0, 2, 3]]],
device=device,
dtype=dtype,
) # 1x8x3
patches = kornia.geometry.transform.crop_and_resize3d(inp, boxes, (depth, height, width))
assert_close(patches, expected)
def test_crop_batch(self, device, dtype):
inp = torch.cat(
[
torch.arange(0.0, 64.0, device=device, dtype=dtype).view(1, 1, 4, 4, 4),
torch.arange(0.0, 128.0, step=2, device=device, dtype=dtype).view(1, 1, 4, 4, 4),
],
dim=0,
)
depth, height, width = 2, 2, 2
expected = torch.tensor(
[
[[[[16.0000, 19.0000], [24.0000, 27.0000]], [[48.0000, 51.0000], [56.0000, 59.0000]]]],
[[[[0.0000, 6.0000], [16.0000, 22.0000]], [[64.0000, 70.0000], [80.0000, 86.0000]]]],
],
device=device,
dtype=dtype,
)
boxes = torch.tensor(
[
[[0, 0, 1], [3, 0, 1], [3, 2, 1], [0, 2, 1], [0, 0, 3], [3, 0, 3], [3, 2, 3], [0, 2, 3]],
[[0, 0, 0], [3, 0, 0], [3, 2, 0], [0, 2, 0], [0, 0, 2], [3, 0, 2], [3, 2, 2], [0, 2, 2]],
],
device=device,
dtype=dtype,
) # 2x8x3
patches = kornia.geometry.transform.crop_and_resize3d(inp, boxes, (depth, height, width), align_corners=True)
assert_close(patches, expected)
def test_gradcheck(self, device, dtype):
img = torch.arange(0.0, 64.0, device=device, dtype=dtype).view(1, 1, 4, 4, 4)
img = utils.tensor_to_gradcheck_var(img) # to var
boxes = torch.tensor(
[[[0, 0, 1], [3, 0, 1], [3, 2, 1], [0, 2, 1], [0, 0, 3], [3, 0, 3], [3, 2, 3], [0, 2, 3]]],
device=device,
dtype=dtype,
) # 1x8x3
boxes = utils.tensor_to_gradcheck_var(boxes, requires_grad=False) # to var
assert gradcheck(kornia.geometry.transform.crop_and_resize3d, (img, boxes, (4, 3, 2)), raise_exception=True)
def test_jit(self, device, dtype):
# Define script
op = kornia.geometry.transform.crop_and_resize3d
op_script = torch.jit.script(op)
img = torch.arange(0.0, 64.0, device=device, dtype=dtype).view(1, 1, 4, 4, 4)
boxes = torch.tensor(
[[[0, 0, 1], [3, 0, 1], [3, 2, 1], [0, 2, 1], [0, 0, 3], [3, 0, 3], [3, 2, 3], [0, 2, 3]]],
device=device,
dtype=dtype,
) # 1x8x3
actual = op_script(img, boxes, (4, 3, 2))
expected = op(img, boxes, (4, 3, 2))
assert_close(actual, expected)
class TestCenterCrop3D:
@pytest.mark.parametrize("crop_size", [(3, 5, 7), (5, 3, 7), (7, 3, 5)])
def test_center_crop_357(self, crop_size, device, dtype):
inp = torch.arange(0.0, 343.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7)
expected = inp[
:,
:,
(inp.size(2) // 2 - crop_size[0] // 2) : (inp.size(2) // 2 + crop_size[0] // 2 + 1),
(inp.size(3) // 2 - crop_size[1] // 2) : (inp.size(3) // 2 + crop_size[1] // 2 + 1),
(inp.size(4) // 2 - crop_size[2] // 2) : (inp.size(4) // 2 + crop_size[2] // 2 + 1),
]
out_crop = kornia.geometry.transform.center_crop3d(inp, crop_size, align_corners=True)
assert_close(out_crop, expected, rtol=1e-4, atol=1e-4)
@pytest.mark.parametrize("crop_size", [(3, 5, 7), (5, 3, 7), (7, 3, 5)])
def test_center_crop_357_batch(self, crop_size, device, dtype):
inp = torch.cat(
[
torch.arange(0.0, 343.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7),
torch.arange(343.0, 686.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7),
]
)
expected = inp[
:,
:,
(inp.size(2) // 2 - crop_size[0] // 2) : (inp.size(2) // 2 + crop_size[0] // 2 + 1),
(inp.size(3) // 2 - crop_size[1] // 2) : (inp.size(3) // 2 + crop_size[1] // 2 + 1),
(inp.size(4) // 2 - crop_size[2] // 2) : (inp.size(4) // 2 + crop_size[2] // 2 + 1),
]
out_crop = kornia.geometry.transform.center_crop3d(inp, crop_size, align_corners=True)
assert_close(out_crop, expected, rtol=1e-4, atol=1e-4)
def test_gradcheck(self, device, dtype):
img = torch.arange(0.0, 343.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7)
img = utils.tensor_to_gradcheck_var(img) # to var
assert gradcheck(kornia.geometry.transform.center_crop3d, (img, (3, 5, 7)), raise_exception=True)
def test_jit(self, device, dtype):
# Define script
op = kornia.geometry.transform.center_crop3d
op_script = torch.jit.script(op)
img = torch.ones(4, 3, 5, 6, 7, device=device, dtype=dtype)
actual = op_script(img, (4, 3, 2))
expected = kornia.geometry.transform.center_crop3d(img, (4, 3, 2))
assert_close(actual, expected, rtol=1e-4, atol=1e-4)
class TestCropByBoxes3D:
def test_crop_by_boxes_no_resizing(self, device, dtype):
inp = torch.arange(0.0, 343.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7)
src_box = torch.tensor(
[
[
[1.0, 1.0, 1.0],
[3.0, 1.0, 1.0],
[3.0, 3.0, 1.0],
[1.0, 3.0, 1.0],
[1.0, 1.0, 2.0],
[3.0, 1.0, 2.0],
[3.0, 3.0, 2.0],
[1.0, 3.0, 2.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
dst_box = torch.tensor(
[
[
[0.0, 0.0, 0.0],
[2.0, 0.0, 0.0],
[2.0, 2.0, 0.0],
[0.0, 2.0, 0.0],
[0.0, 0.0, 1.0],
[2.0, 0.0, 1.0],
[2.0, 2.0, 1.0],
[0.0, 2.0, 1.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
expected = inp[:, :, 1:3, 1:4, 1:4]
patches = kornia.geometry.transform.crop_by_boxes3d(inp, src_box, dst_box, align_corners=True)
assert_close(patches, expected, rtol=1e-4, atol=1e-4)
def test_crop_by_boxes_resizing(self, device, dtype):
inp = torch.arange(0.0, 343.0, device=device, dtype=dtype).view(1, 1, 7, 7, 7)
src_box = torch.tensor(
[
[
[1.0, 1.0, 1.0],
[3.0, 1.0, 1.0],
[3.0, 3.0, 1.0],
[1.0, 3.0, 1.0],
[1.0, 1.0, 2.0],
[3.0, 1.0, 2.0],
[3.0, 3.0, 2.0],
[1.0, 3.0, 2.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
dst_box = torch.tensor(
[
[
[0.0, 0.0, 0.0],
[1.0, 0.0, 0.0],
[1.0, 1.0, 0.0],
[0.0, 1.0, 0.0],
[0.0, 0.0, 1.0],
[1.0, 0.0, 1.0],
[1.0, 1.0, 1.0],
[0.0, 1.0, 1.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
expected = torch.tensor(
[[[[[57.0000, 59.0000], [71.0000, 73.0000]], [[106.0000, 108.0000], [120.0000, 122.0000]]]]],
device=device,
dtype=dtype,
)
patches = kornia.geometry.transform.crop_by_boxes3d(inp, src_box, dst_box, align_corners=True)
assert_close(patches, expected, rtol=1e-4, atol=1e-4)
def test_jit(self, device, dtype):
# Define script
op = kornia.geometry.transform.crop_by_boxes3d
op_script = torch.jit.script(op)
# Define input
inp = torch.randn((1, 1, 7, 7, 7), device=device, dtype=dtype)
src_box = torch.tensor(
[
[
[1.0, 1.0, 1.0],
[3.0, 1.0, 1.0],
[3.0, 3.0, 1.0],
[1.0, 3.0, 1.0],
[1.0, 1.0, 2.0],
[3.0, 1.0, 2.0],
[3.0, 3.0, 2.0],
[1.0, 3.0, 2.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
dst_box = torch.tensor(
[
[
[0.0, 0.0, 0.0],
[1.0, 0.0, 0.0],
[1.0, 1.0, 0.0],
[0.0, 1.0, 0.0],
[0.0, 0.0, 1.0],
[1.0, 0.0, 1.0],
[1.0, 1.0, 1.0],
[0.0, 1.0, 1.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
actual = op_script(inp, src_box, dst_box, align_corners=True)
expected = op(inp, src_box, dst_box, align_corners=True)
assert_close(actual, expected, rtol=1e-4, atol=1e-4)
def test_gradcheck(self, device, dtype):
inp = torch.randn((1, 1, 7, 7, 7), device=device, dtype=dtype)
src_box = torch.tensor(
[
[
[1.0, 1.0, 1.0],
[3.0, 1.0, 1.0],
[3.0, 3.0, 1.0],
[1.0, 3.0, 1.0],
[1.0, 1.0, 2.0],
[3.0, 1.0, 2.0],
[3.0, 3.0, 2.0],
[1.0, 3.0, 2.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
dst_box = torch.tensor(
[
[
[0.0, 0.0, 0.0],
[1.0, 0.0, 0.0],
[1.0, 1.0, 0.0],
[0.0, 1.0, 0.0],
[0.0, 0.0, 1.0],
[1.0, 0.0, 1.0],
[1.0, 1.0, 1.0],
[0.0, 1.0, 1.0],
]
],
device=device,
dtype=dtype,
) # 1x8x3
inp = utils.tensor_to_gradcheck_var(inp, requires_grad=True) # to var
assert gradcheck(kornia.geometry.transform.crop_by_boxes3d, (inp, src_box, dst_box), raise_exception=True)