fake_quantize.py 1015 B

1234567891011121314151617181920212223242526272829303132
  1. # flake8: noqa: F401
  2. r"""
  3. This file is in the process of migration to `torch/ao/quantization`, and
  4. is kept here for compatibility while the migration process is ongoing.
  5. If you are adding a new entry/functionality, please, add it to the
  6. `torch/ao/quantization/fake_quantize.py`, while adding an import statement
  7. here.
  8. """
  9. from torch.ao.quantization.fake_quantize import (
  10. _is_per_channel,
  11. _is_per_tensor,
  12. _is_symmetric_quant,
  13. FakeQuantizeBase,
  14. FakeQuantize,
  15. FixedQParamsFakeQuantize,
  16. FusedMovingAvgObsFakeQuantize,
  17. default_fake_quant,
  18. default_weight_fake_quant,
  19. default_fixed_qparams_range_neg1to1_fake_quant,
  20. default_fixed_qparams_range_0to1_fake_quant,
  21. default_per_channel_weight_fake_quant,
  22. default_histogram_fake_quant,
  23. default_fused_act_fake_quant,
  24. default_fused_wt_fake_quant,
  25. default_fused_per_channel_wt_fake_quant,
  26. _is_fake_quant_script_module,
  27. disable_fake_quant,
  28. enable_fake_quant,
  29. disable_observer,
  30. enable_observer,
  31. )