123456789101112131415161718 |
- # flake8: noqa: F401
- r"""Quantized Modules
- This file is in the process of migration to `torch/ao/nn/quantized`, and
- is kept here for compatibility while the migration process is ongoing.
- If you are adding a new entry/functionality, please, add it to the
- appropriate file under the `torch/ao/nn/quantized/modules`,
- while adding an import statement here.
- """
- from torch.ao.nn.quantized.modules.activation import ELU
- from torch.ao.nn.quantized.modules.activation import Hardswish
- from torch.ao.nn.quantized.modules.activation import LeakyReLU
- from torch.ao.nn.quantized.modules.activation import MultiheadAttention
- from torch.ao.nn.quantized.modules.activation import PReLU
- from torch.ao.nn.quantized.modules.activation import ReLU6
- from torch.ao.nn.quantized.modules.activation import Sigmoid
- from torch.ao.nn.quantized.modules.activation import Softmax
|