| # flake8: noqa: F401 | |
| r""" | |
| Utils shared by different modes of quantization (eager/graph) | |
| This file is in the process of migration to `torch/ao/quantization`, and | |
| is kept here for compatibility while the migration process is ongoing. | |
| If you are adding a new entry/functionality, please, add it to the | |
| `torch/ao/quantization/utils.py`, while adding an import statement | |
| here. | |
| """ | |
| from torch.ao.quantization.utils import ( | |
| activation_dtype, | |
| activation_is_int8_quantized, | |
| activation_is_statically_quantized, | |
| calculate_qmin_qmax, | |
| check_min_max_valid, | |
| get_combined_dict, | |
| get_qconfig_dtypes, | |
| get_qparam_dict, | |
| get_quant_type, | |
| get_swapped_custom_module_class, | |
| getattr_from_fqn, | |
| is_per_channel, | |
| is_per_tensor, | |
| weight_dtype, | |
| weight_is_quantized, | |
| weight_is_statically_quantized, | |
| ) | |