1234567891011121314151617181920212223242526272829 |
- # flake8: noqa: F401
- r"""
- Utils shared by different modes of quantization (eager/graph)
- This file is in the process of migration to `torch/ao/quantization`, and
- is kept here for compatibility while the migration process is ongoing.
- If you are adding a new entry/functionality, please, add it to the
- `torch/ao/quantization/utils.py`, while adding an import statement
- here.
- """
- from torch.ao.quantization.utils import (
- activation_dtype,
- activation_is_int8_quantized,
- activation_is_statically_quantized,
- calculate_qmin_qmax,
- check_min_max_valid,
- get_combined_dict,
- get_qconfig_dtypes,
- get_qparam_dict,
- get_quant_type,
- get_swapped_custom_module_class,
- getattr_from_fqn,
- is_per_channel,
- is_per_tensor,
- weight_dtype,
- weight_is_quantized,
- weight_is_statically_quantized,
- )
|