| # flake8: noqa: F401 |
| r""" |
| Utils shared by different modes of quantization (eager/graph) |
| |
| This file is in the process of migration to `torch/ao/quantization`, and |
| is kept here for compatibility while the migration process is ongoing. |
| If you are adding a new entry/functionality, please, add it to the |
| `torch/ao/quantization/utils.py`, while adding an import statement |
| here. |
| """ |
| |
| from torch.ao.quantization.utils import ( |
| activation_dtype, |
| activation_is_int8_quantized, |
| activation_is_statically_quantized, |
| calculate_qmin_qmax, |
| check_min_max_valid, |
| get_combined_dict, |
| get_qconfig_dtypes, |
| get_qparam_dict, |
| get_quant_type, |
| get_swapped_custom_module_class, |
| getattr_from_fqn, |
| is_per_channel, |
| is_per_tensor, |
| weight_dtype, |
| weight_is_quantized, |
| weight_is_statically_quantized, |
| ) |