|
|
import os |
|
|
os.environ["CUDA_VISIBLE_DEVICES"] = "-1" |
|
|
os.environ["JAX_PLATFORMS"] = "cpu" |
|
|
|
|
|
import pytest |
|
|
import jax |
|
|
import jax.numpy as jnp |
|
|
import numpy as np |
|
|
import torch |
|
|
|
|
|
from quantization import quantization_types |
|
|
|
|
|
|
|
|
torch.autograd.grad_mode.set_grad_enabled(False) |
|
|
backends = {"jax": jnp, "numpy": np, "torch": torch} |
|
|
scales = (0.5, 1.0, 1.5) |
|
|
|
|
|
|
|
|
@pytest.fixture(params=backends.values(), ids=backends.keys()) |
|
|
def backend(request): |
|
|
"""Fixture to parameterize over all backends.""" |
|
|
return request.param |
|
|
|
|
|
|
|
|
def _uint8_range(backend): |
|
|
"""Fixture to provide uint8 range for each backend.""" |
|
|
return backend.arange(256, dtype=backend.uint8) |
|
|
|
|
|
|
|
|
uint8_range = pytest.fixture(_uint8_range) |
|
|
|
|
|
|
|
|
def _float_range(backend): |
|
|
"""Fixture to provide float range for each backend.""" |
|
|
return backend.linspace(-4, 4, 101, dtype=backend.float32) |
|
|
|
|
|
|
|
|
float_range = pytest.fixture(_float_range) |
|
|
|
|
|
|
|
|
@pytest.fixture(params=quantization_types) |
|
|
def cls(request): |
|
|
"""Fixture to parameterize over all quantization classes.""" |
|
|
return request.param |
|
|
|
|
|
|
|
|
@pytest.fixture |
|
|
def default_instance(cls): |
|
|
"""Fixture to parameterize over default instances of all quantization classes.""" |
|
|
return cls() |
|
|
|
|
|
|
|
|
@pytest.fixture(params=scales) |
|
|
def scale(request): |
|
|
"""Fixture to parameterize over different scale values.""" |
|
|
return request.param |
|
|
|
|
|
|
|
|
@pytest.fixture |
|
|
def instance(cls, scale): |
|
|
"""Fixture to parameterize over instances of all quantization classes, swept over all scale values.""" |
|
|
return cls(scale=scale) |
|
|
|
|
|
|
|
|
def test_interface(cls, default_instance): |
|
|
"""Test that all quantization classes and instances have the expected methods.""" |
|
|
assert isinstance(default_instance, cls) |
|
|
classattrs = () |
|
|
classmethods = () |
|
|
instanceattrs = ("scale",) |
|
|
instancemethods = ("quantize", "dequantize", "nonlinearity", "inv_nonlinearity") |
|
|
interface = { |
|
|
cls: (classattrs, classmethods), |
|
|
default_instance: (instanceattrs, instancemethods) |
|
|
} |
|
|
for obj, (attrs, methods) in interface.items(): |
|
|
for attr in attrs + methods: |
|
|
assert hasattr(obj, attr) |
|
|
for method in methods: |
|
|
assert callable(getattr(obj, method)) |
|
|
|
|
|
|
|
|
def test_quantize_returns_uint8(default_instance, float_range): |
|
|
"""Test that quantize returns uint8 values.""" |
|
|
quantized = default_instance.quantize(float_range) |
|
|
assert "uint8" in str(quantized.dtype).lower() |
|
|
|
|
|
|
|
|
def test_dequantize_returns_float32(default_instance, uint8_range): |
|
|
"""Test that dequantize returns float32 values.""" |
|
|
dequantized = default_instance.dequantize(uint8_range) |
|
|
assert "float32" in str(dequantized.dtype).lower() |
|
|
|
|
|
|
|
|
def test_roundtrip_consistency(instance, uint8_range): |
|
|
"""Test that dequantize->quantize is lossless.""" |
|
|
dequantized = instance.dequantize(uint8_range) |
|
|
requantized = instance.quantize(dequantized) |
|
|
np.testing.assert_array_equal(uint8_range, requantized) |
|
|
|
|
|
|
|
|
def test_cross_backend_consistency(instance): |
|
|
"""Test that dequantization gives identical results across all backends.""" |
|
|
results = { |
|
|
name: instance.dequantize(_uint8_range(backend)) |
|
|
for name, backend in backends.items() |
|
|
} |
|
|
baseline_name = list(backends.keys())[0] |
|
|
baseline_result = results[baseline_name] |
|
|
for name, result in results.items(): |
|
|
if name == baseline_name: |
|
|
continue |
|
|
np.testing.assert_array_almost_equal( |
|
|
result, |
|
|
baseline_result, |
|
|
err_msg=f"{name} result doesn't match {baseline_name}", |
|
|
) |
|
|
|
|
|
|
|
|
def test_jax_jit(instance): |
|
|
"""Test that the JAX backend works when the quantization methods are jitted.""" |
|
|
data = _uint8_range(jnp) |
|
|
|
|
|
def roundtrip(x): |
|
|
return instance.quantize(instance.dequantize(x)) |
|
|
|
|
|
nojit = roundtrip(data) |
|
|
yesjit = jax.jit(roundtrip)(data) |
|
|
np.testing.assert_array_equal(yesjit, nojit) |
|
|
|
|
|
|
|
|
def test_torch_compile(instance): |
|
|
"""Test that the PyTorch backend works when the quantization methods are compiled.""" |
|
|
data = _uint8_range(torch) |
|
|
|
|
|
def roundtrip(x): |
|
|
return instance.quantize(instance.dequantize(x)) |
|
|
|
|
|
nojit = roundtrip(data) |
|
|
yesjit = torch.compile(roundtrip)(data) |
|
|
np.testing.assert_array_equal(yesjit, nojit) |
|
|
|