# Copyright 2020 MONAI Consortium # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest import numpy as np import torch from parameterized import parameterized from monai.metrics import DiceMetric, compute_meandice # keep background TEST_CASE_1 = [ # y (1, 1, 2, 2), y_pred (1, 1, 2, 2), expected out (1, 1) { "y_pred": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]), "y": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]), "include_background": True, "to_onehot_y": False, "mutually_exclusive": False, "logit_thresh": 0.5, "sigmoid": True, }, [[0.8]], ] # remove background and not One-Hot target TEST_CASE_2 = [ # y (2, 1, 2, 2), y_pred (2, 3, 2, 2), expected out (2, 2) (no background) { "y_pred": torch.tensor( [ [[[-1.0, 3.0], [2.0, -4.0]], [[0.0, -1.0], [3.0, 2.0]], [[0.0, 1.0], [2.0, -1.0]]], [[[-2.0, 0.0], [3.0, 1.0]], [[0.0, 2.0], [1.0, -2.0]], [[-1.0, 2.0], [4.0, 0.0]]], ] ), "y": torch.tensor([[[[1.0, 2.0], [1.0, 0.0]]], [[[1.0, 1.0], [2.0, 0.0]]]]), "include_background": False, "to_onehot_y": True, "mutually_exclusive": True, }, [[0.5000, 0.0000], [0.6666, 0.6666]], ] # should return Nan for all labels=0 case and skip for MeanDice TEST_CASE_3 = [ { "y_pred": torch.zeros(2, 3, 2, 2), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[1.0, 0.0], [0.0, 1.0]]]]), "include_background": True, "to_onehot_y": True, "mutually_exclusive": True, }, [[False, True, True], [False, False, True]], ] TEST_CASE_4 = [ {"include_background": True, "to_onehot_y": True, "reduction": "mean_batch"}, { "y_pred": torch.tensor( [ [[[1.0, 1.0], [1.0, 0.0]], [[0.0, 1.0], [0.0, 0.0]], [[0.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 0.0]]], ] ), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[1.0, 1.0], [2.0, 0.0]]]]), }, [0.6786, 0.4000, 0.6667], ] TEST_CASE_5 = [ {"include_background": True, "to_onehot_y": True, "reduction": "mean"}, { "y_pred": torch.tensor( [ [[[1.0, 1.0], [1.0, 0.0]], [[0.0, 1.0], [0.0, 0.0]], [[0.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 0.0]]], ] ), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[1.0, 1.0], [2.0, 0.0]]]]), }, 0.689683, ] TEST_CASE_6 = [ {"include_background": True, "to_onehot_y": True, "reduction": "sum_batch"}, { "y_pred": torch.tensor( [ [[[1.0, 1.0], [1.0, 0.0]], [[0.0, 1.0], [0.0, 0.0]], [[0.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 0.0]]], ] ), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[0.0, 0.0], [0.0, 0.0]]]]), }, [1.7143, 0.0000, 0.0000], ] TEST_CASE_7 = [ {"include_background": True, "to_onehot_y": True, "reduction": "mean"}, { "y_pred": torch.tensor( [ [[[1.0, 1.0], [1.0, 0.0]], [[0.0, 1.0], [0.0, 0.0]], [[0.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 0.0]]], ] ), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[0.0, 0.0], [0.0, 0.0]]]]), }, 0.857143, ] TEST_CASE_8 = [ {"to_onehot_y": True, "include_background": False, "reduction": "sum_batch"}, { "y_pred": torch.tensor( [ [[[1.0, 1.0], [1.0, 0.0]], [[0.0, 1.0], [0.0, 0.0]], [[0.0, 1.0], [1.0, 1.0]]], [[[1.0, 0.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 1.0]], [[0.0, 1.0], [1.0, 0.0]]], ] ), "y": torch.tensor([[[[0.0, 0.0], [0.0, 0.0]]], [[[0.0, 0.0], [0.0, 0.0]]]]), }, [0.0000, 0.0000], ] TEST_CASE_9 = [ {"y": torch.from_numpy(np.ones((2, 2, 3, 3))), "y_pred": torch.from_numpy(np.ones((2, 2, 3, 3)))}, [[1.0000, 1.0000], [1.0000, 1.0000]], ] TEST_CASE_10 = [ # y (1, 1, 2, 2), y_pred (1, 1, 2, 2), expected out (1, 1) { "y_pred": torch.tensor([[[[1.0, -1.0], [-1.0, 1.0]]]]), "y": torch.tensor([[[[1.0, 0.0], [1.0, 1.0]]]]), "include_background": True, "to_onehot_y": False, "mutually_exclusive": False, "logit_thresh": 0.0, "other_act": torch.tanh, }, [[0.8]], ] class TestComputeMeanDice(unittest.TestCase): @parameterized.expand([TEST_CASE_1, TEST_CASE_2, TEST_CASE_9, TEST_CASE_10]) def test_value(self, input_data, expected_value): result = compute_meandice(**input_data) np.testing.assert_allclose(result.cpu().numpy(), expected_value, atol=1e-4) @parameterized.expand([TEST_CASE_3]) def test_nans(self, input_data, expected_value): result = compute_meandice(**input_data) self.assertTrue(np.allclose(np.isnan(result.cpu().numpy()), expected_value)) # DiceMetric class tests @parameterized.expand([TEST_CASE_1, TEST_CASE_2]) def test_value_class(self, input_data, expected_value): # same test as for compute_meandice vals = dict() vals["y_pred"] = input_data.pop("y_pred") vals["y"] = input_data.pop("y") dice_metric = DiceMetric(**input_data, reduction="none") result = dice_metric(**vals) np.testing.assert_allclose(result.cpu().numpy(), expected_value, atol=1e-4) @parameterized.expand([TEST_CASE_4, TEST_CASE_5, TEST_CASE_6, TEST_CASE_7, TEST_CASE_8]) def test_nans_class(self, params, input_data, expected_value): dice_metric = DiceMetric(**params) result = dice_metric(**input_data) np.testing.assert_allclose(result.cpu().numpy(), expected_value, atol=1e-4) if __name__ == "__main__": unittest.main()