| # flake8: noqa: F401 |
| r"""Quantized Modules |
| |
| This file is in the process of migration to `torch/ao/nn/quantized`, and |
| is kept here for compatibility while the migration process is ongoing. |
| If you are adding a new entry/functionality, please, add it to the |
| appropriate file under the `torch/ao/nn/quantized/modules`, |
| while adding an import statement here. |
| """ |
| |
| from torch.ao.nn.quantized.modules.activation import ELU |
| from torch.ao.nn.quantized.modules.activation import Hardswish |
| from torch.ao.nn.quantized.modules.activation import LeakyReLU |
| from torch.ao.nn.quantized.modules.activation import MultiheadAttention |
| from torch.ao.nn.quantized.modules.activation import PReLU |
| from torch.ao.nn.quantized.modules.activation import ReLU6 |
| from torch.ao.nn.quantized.modules.activation import Sigmoid |
| from torch.ao.nn.quantized.modules.activation import Softmax |