| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| """Tests for `utils.py`.""" |
|
|
| from unittest import mock |
|
|
| from absl.testing import absltest |
| from absl.testing import parameterized |
| import chex |
| import jax |
| import jax.numpy as jnp |
| import numpy as np |
| from optax._src import alias |
| from optax._src import combine |
| from optax._src import linesearch |
| from optax._src import transform |
| from optax._src import update |
| from optax._src import utils |
|
|
|
|
| def _shape_to_tuple(shape): |
| if isinstance(shape, tuple): |
| return shape |
| return tuple([shape]) |
|
|
|
|
| class ScaleGradientTest(parameterized.TestCase): |
|
|
| @parameterized.product( |
| inputs=[-1.0, 0.0, 1.0], scale=[-0.5, 0.0, 0.5, 1.0, 2.0] |
| ) |
| @mock.patch.object(jax.lax, 'stop_gradient', wraps=jax.lax.stop_gradient) |
| def test_scale_gradient(self, mock_sg, inputs, scale): |
| def fn(inputs): |
| outputs = utils.scale_gradient(inputs, scale) |
| return outputs**2 |
|
|
| grad = jax.grad(fn) |
| self.assertEqual(grad(inputs), 2 * inputs * scale) |
| if scale == 0.0: |
| mock_sg.assert_called_once_with(inputs) |
| else: |
| self.assertFalse(mock_sg.called) |
| self.assertEqual(fn(inputs), inputs**2) |
|
|
| @parameterized.product(scale=[-0.5, 0.0, 0.5, 1.0, 2.0]) |
| def test_scale_gradient_pytree(self, scale): |
| def fn(inputs): |
| outputs = utils.scale_gradient(inputs, scale) |
| outputs = jax.tree_util.tree_map(lambda x: x**2, outputs) |
| return sum(jax.tree_util.tree_leaves(outputs)) |
|
|
| inputs = dict(a=-1.0, b=dict(c=(2.0,), d=0.0)) |
|
|
| grad = jax.grad(fn) |
| grads = grad(inputs) |
| jax.tree_util.tree_map( |
| lambda i, g: self.assertEqual(g, 2 * i * scale), inputs, grads |
| ) |
| self.assertEqual( |
| fn(inputs), |
| sum( |
| jax.tree_util.tree_leaves( |
| jax.tree_util.tree_map(lambda x: x**2, inputs) |
| ) |
| ), |
| ) |
|
|
|
|
| class MultiNormalDiagFromLogScaleTest(parameterized.TestCase): |
|
|
| def _get_loc_scale(self, loc_shape, scale_shape): |
| loc = 1.5 * jnp.ones(shape=loc_shape, dtype=jnp.float32) |
| scale = 0.5 * jnp.ones(shape=scale_shape, dtype=jnp.float32) |
| return loc, scale |
|
|
| @parameterized.parameters( |
| (1, 1, 1), |
| (5, 5, 5), |
| ((2, 3), (2, 3), (2, 3)), |
| ((1, 4), (3, 4), (3, 4)), |
| ((1, 2, 1, 3), (2, 1, 4, 3), (2, 2, 4, 3)), |
| ) |
| def test_init_successful_broadcast( |
| self, loc_shape, scale_shape, broadcasted_shape |
| ): |
| loc, scale = self._get_loc_scale(loc_shape, scale_shape) |
| dist = utils.multi_normal(loc, scale) |
| self.assertIsInstance(dist, utils.MultiNormalDiagFromLogScale) |
| mean, log_scale = dist.params |
| self.assertEqual(tuple(mean.shape), _shape_to_tuple(loc_shape)) |
| self.assertEqual(tuple(log_scale.shape), _shape_to_tuple(scale_shape)) |
| self.assertEqual( |
| tuple(dist._param_shape), _shape_to_tuple(broadcasted_shape) |
| ) |
|
|
| @parameterized.parameters( |
| (2, 3), |
| ((2, 3), (3, 2)), |
| ((2, 4), (3, 4)), |
| ((1, 2, 1, 3), (2, 1, 4, 4)), |
| ) |
| def test_init_unsuccessful_broadcast(self, loc_shape, scale_shape): |
| loc, scale = self._get_loc_scale(loc_shape, scale_shape) |
| with self.assertRaisesRegex( |
| ValueError, 'Incompatible shapes for broadcasting' |
| ): |
| utils.multi_normal(loc, scale) |
|
|
| @parameterized.parameters(list, tuple) |
| def test_sample_input_sequence_types(self, sample_type): |
| sample_shape = sample_type((4, 5)) |
| loc_shape = scale_shape = (2, 3) |
| loc, scale = self._get_loc_scale(loc_shape, scale_shape) |
| dist = utils.multi_normal(loc, scale) |
| samples = dist.sample(sample_shape, jax.random.PRNGKey(239)) |
| self.assertEqual(samples.shape, tuple(sample_shape) + loc_shape) |
|
|
| @parameterized.named_parameters([ |
| ('1d', 1), |
| ('2d', (2, 3)), |
| ('4d', (1, 2, 3, 4)), |
| ]) |
| def test_log_prob(self, shape): |
| loc, scale = self._get_loc_scale(shape, shape) |
| dist = utils.multi_normal(loc, scale) |
| probs = dist.log_prob(jnp.ones(shape=shape, dtype=jnp.float32)) |
| self.assertEqual(probs.shape, ()) |
|
|
|
|
| class HelpersTest(chex.TestCase): |
|
|
| @parameterized.parameters([ |
| (1, 1), |
| (3, 3), |
| (1, 3), |
| (2, 3), |
| ]) |
| def test_set_diags_valid(self, n, d): |
| def _all_but_diag(matrix): |
| return matrix - jnp.diag(jnp.diag(matrix)) |
|
|
| a = jnp.ones(shape=(n, d, d)) * 10 |
| new_diags = jnp.arange(n * d).reshape((n, d)) |
| res = utils.set_diags(a, new_diags) |
| for i in range(n): |
| np.testing.assert_array_equal(jnp.diag(res[i]), new_diags[i]) |
| np.testing.assert_array_equal(_all_but_diag(res[i]), _all_but_diag(a[i])) |
|
|
| @parameterized.named_parameters([ |
| ('1d', 1), |
| ('2d', (2, 3)), |
| ('4d', (1, 2, 3, 4)), |
| ]) |
| def test_set_diag_a_raises(self, a_shape): |
| a = jnp.ones(shape=a_shape) |
| new_diags = jnp.zeros(shape=(2, 2)) |
| with self.assertRaisesRegex(ValueError, 'Expected `a` to be a 3D tensor'): |
| utils.set_diags(a, new_diags) |
|
|
| @parameterized.named_parameters([ |
| ('1d', 1), |
| ('3d', (2, 3, 4)), |
| ('4d', (1, 2, 3, 4)), |
| ]) |
| def test_set_diag_new_diags_raises(self, new_diags_shape): |
| a = jnp.ones(shape=(3, 2, 2)) |
| new_diags = jnp.zeros(shape=new_diags_shape) |
| with self.assertRaisesRegex( |
| ValueError, 'Expected `new_diags` to be a 2D array' |
| ): |
| utils.set_diags(a, new_diags) |
|
|
| @parameterized.parameters([ |
| (1, 1, 2), |
| (3, 3, 4), |
| (1, 3, 5), |
| (2, 3, 2), |
| ]) |
| def test_set_diag_a_shape_mismatch_raises(self, n, d, d1): |
| a = jnp.ones(shape=(n, d, d1)) |
| new_diags = jnp.zeros(shape=(n, d)) |
| with self.assertRaisesRegex( |
| ValueError, 'Shape mismatch: expected `a.shape`' |
| ): |
| utils.set_diags(a, new_diags) |
|
|
| @parameterized.parameters([ |
| (1, 1, 1, 3), |
| (3, 3, 4, 3), |
| (1, 3, 1, 5), |
| (2, 3, 6, 7), |
| ]) |
| def test_set_diag_new_diags_shape_mismatch_raises(self, n, d, n1, d1): |
| a = jnp.ones(shape=(n, d, d)) |
| new_diags = jnp.zeros(shape=(n1, d1)) |
| with self.assertRaisesRegex( |
| ValueError, 'Shape mismatch: expected `new_diags.shape`' |
| ): |
| utils.set_diags(a, new_diags) |
|
|
| @parameterized.named_parameters([ |
| ('1d-single', 1), |
| ('1d', 10), |
| ('2d', (1, 2)), |
| ('3d', (10, 3, 2)), |
| ('4d', (2, 3, 4, 5)), |
| ('6d', (1, 2, 3, 4, 5, 6)), |
| ('8d', (5, 4, 7, 6, 1, 2, 3, 1)), |
| ]) |
| def test_tile_second_to_last_dim(self, shape): |
| shape = _shape_to_tuple(shape) |
| elems = jnp.prod(jnp.array(shape)) |
| matrix = jnp.arange(elems).reshape(shape) |
| result = utils.tile_second_to_last_dim(matrix) |
| self.assertEqual(result.shape, shape + (shape[-1],)) |
| np.testing.assert_array_equal(result[..., -1], matrix) |
| np.testing.assert_array_equal(result[..., 0], matrix) |
|
|
| @parameterized.parameters([ |
| (None, None), |
| (jnp.float32, np.dtype('float32')), |
| (jnp.int32, np.dtype('int32')), |
| (jnp.bfloat16, np.dtype('bfloat16')), |
| ]) |
| def test_canonicalize_dtype(self, dtype, expected_dtype): |
| canonical = utils.canonicalize_dtype(dtype) |
| self.assertIs(canonical, expected_dtype) |
|
|
| @chex.variants( |
| with_jit=True, |
| without_jit=True, |
| with_pmap=False, |
| with_device=True, |
| without_device=True, |
| ) |
| def test_value_and_grad_from_state(self): |
| def fn(x): |
| return jnp.sum(x**2) |
|
|
| value_and_grad_ = utils.value_and_grad_from_state(fn) |
|
|
| value_and_grad = self.variant(value_and_grad_) |
|
|
| params = jnp.array([1.0, 2.0, 3.0]) |
|
|
| |
| opt = transform.scale_by_adam() |
| state = opt.init(params) |
| self.assertRaises(ValueError, value_and_grad, params, state=state) |
|
|
| |
| opt = combine.chain( |
| linesearch.scale_by_backtracking_linesearch(max_backtracking_steps=15), |
| linesearch.scale_by_backtracking_linesearch(max_backtracking_steps=15), |
| ) |
| state = opt.init(params) |
| self.assertRaises(KeyError, value_and_grad, params, state=state) |
|
|
| |
| opt = combine.chain( |
| alias.sgd(learning_rate=1.0), |
| linesearch.scale_by_backtracking_linesearch( |
| max_backtracking_steps=15, store_grad=True |
| ), |
| ) |
| state = opt.init(params) |
| value, grad = value_and_grad(params, state=state) |
| updates, state = opt.update( |
| grad, state, params, value=value, grad=grad, value_fn=fn |
| ) |
| params = update.apply_updates(params, updates) |
| params = jax.block_until_ready(params) |
|
|
| def false_fn(_): |
| return 1.0 |
|
|
| false_value_and_grad_ = utils.value_and_grad_from_state(false_fn) |
| false_value_and_grad = self.variant(false_value_and_grad_) |
|
|
| |
| |
| value, _ = false_value_and_grad(params, state=state) |
| self.assertNotEqual(value, 1.0) |
|
|
| def test_extract_fns_kwargs(self): |
| def fn1(a, b): |
| return a + b |
|
|
| def fn2(c, d): |
| return c + d |
|
|
| kwargs = {'b': 1.0, 'd': 2.0, 'e': 3.0} |
| fns_kwargs, remaining_kwargs = utils._extract_fns_kwargs((fn1, fn2), kwargs) |
| self.assertEqual(fns_kwargs, [{'b': 1.0}, {'d': 2.0}]) |
| self.assertEqual(remaining_kwargs, {'e': 3.0}) |
|
|
|
|
| if __name__ == '__main__': |
| absltest.main() |
|
|