| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | import unittest |
| |
|
| | import numpy as np |
| | import torch |
| |
|
| | from .utils_summarization import build_mask, compute_token_type_ids, process_story, truncate_or_pad |
| |
|
| |
|
| | class SummarizationDataProcessingTest(unittest.TestCase): |
| | def setUp(self): |
| | self.block_size = 10 |
| |
|
| | def test_fit_to_block_sequence_too_small(self): |
| | """Pad the sequence with 0 if the sequence is smaller than the block size.""" |
| | sequence = [1, 2, 3, 4] |
| | expected_output = [1, 2, 3, 4, 0, 0, 0, 0, 0, 0] |
| | self.assertEqual(truncate_or_pad(sequence, self.block_size, 0), expected_output) |
| |
|
| | def test_fit_to_block_sequence_fit_exactly(self): |
| | """Do nothing if the sequence is the right size.""" |
| | sequence = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] |
| | expected_output = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] |
| | self.assertEqual(truncate_or_pad(sequence, self.block_size, 0), expected_output) |
| |
|
| | def test_fit_to_block_sequence_too_big(self): |
| | """Truncate the sequence if it is too long.""" |
| | sequence = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13] |
| | expected_output = [1, 2, 3, 4, 5, 6, 7, 8, 9, 10] |
| | self.assertEqual(truncate_or_pad(sequence, self.block_size, 0), expected_output) |
| |
|
| | def test_process_story_no_highlights(self): |
| | """Processing a story with no highlights returns an empty list for the summary.""" |
| | raw_story = """It was the year of Our Lord one thousand seven hundred and |
| | seventy-five.\n\nSpiritual revelations were conceded to England at that |
| | favoured period, as at this.""" |
| | _, summary_lines = process_story(raw_story) |
| | self.assertEqual(summary_lines, []) |
| |
|
| | def test_process_empty_story(self): |
| | """An empty story returns an empty collection of lines.""" |
| | raw_story = "" |
| | story_lines, summary_lines = process_story(raw_story) |
| | self.assertEqual(story_lines, []) |
| | self.assertEqual(summary_lines, []) |
| |
|
| | def test_process_story_with_missing_period(self): |
| | raw_story = ( |
| | "It was the year of Our Lord one thousand seven hundred and " |
| | "seventy-five\n\nSpiritual revelations were conceded to England " |
| | "at that favoured period, as at this.\n@highlight\n\nIt was the best of times" |
| | ) |
| | story_lines, summary_lines = process_story(raw_story) |
| |
|
| | expected_story_lines = [ |
| | "It was the year of Our Lord one thousand seven hundred and seventy-five.", |
| | "Spiritual revelations were conceded to England at that favoured period, as at this.", |
| | ] |
| | self.assertEqual(expected_story_lines, story_lines) |
| |
|
| | expected_summary_lines = ["It was the best of times."] |
| | self.assertEqual(expected_summary_lines, summary_lines) |
| |
|
| | def test_build_mask_no_padding(self): |
| | sequence = torch.tensor([1, 2, 3, 4]) |
| | expected = torch.tensor([1, 1, 1, 1]) |
| | np.testing.assert_array_equal(build_mask(sequence, 0).numpy(), expected.numpy()) |
| |
|
| | def test_build_mask(self): |
| | sequence = torch.tensor([1, 2, 3, 4, 23, 23, 23]) |
| | expected = torch.tensor([1, 1, 1, 1, 0, 0, 0]) |
| | np.testing.assert_array_equal(build_mask(sequence, 23).numpy(), expected.numpy()) |
| |
|
| | def test_build_mask_with_padding_equal_to_one(self): |
| | sequence = torch.tensor([8, 2, 3, 4, 1, 1, 1]) |
| | expected = torch.tensor([1, 1, 1, 1, 0, 0, 0]) |
| | np.testing.assert_array_equal(build_mask(sequence, 1).numpy(), expected.numpy()) |
| |
|
| | def test_compute_token_type_ids(self): |
| | separator = 101 |
| | batch = torch.tensor([[1, 2, 3, 4, 5, 6], [1, 2, 3, 101, 5, 6], [1, 101, 3, 4, 101, 6]]) |
| | expected = torch.tensor([[1, 1, 1, 1, 1, 1], [1, 1, 1, 0, 0, 0], [1, 0, 0, 0, 1, 1]]) |
| |
|
| | result = compute_token_type_ids(batch, separator) |
| | np.testing.assert_array_equal(result, expected) |
| |
|