Files
neuralchen-SimSwap/PReLU.patch
T
chenxuanhong 3ee304b0e1 update
2021-06-09 13:12:21 +08:00

37 lines
1.3 KiB
Diff

--- /usr/local/lib/python3.5/dist-packages/torch/nn/modules/activation.py
+++ /usr/local/lib/python3.5/dist-packages/torch/nn/modules/activation.py
@@ -37,9 +37,10 @@
- Output: :math:`(N, *)`, same shape as the input
Attributes:
- weight (Tensor): the learnable weights of shape (:attr:`num_parameters`).
+ weight (Tensor): the learnable weights of shape (attr:`num_parameters`).
+ The attr:`dtype` is default to
- .. image:: ../scripts/activation_images/PReLU.png
+ .. image:: scripts/activation_images/PReLU.png
Examples::
@@ -47,17 +48,16 @@
>>> input = torch.randn(2)
>>> output = m(input)
"""
- __constants__ = ['num_parameters']
- num_parameters: int
- def __init__(self, num_parameters: int = 1, init: float = 0.25) -> None:
+ def __init__(self, num_parameters=1, init=0.25):
self.num_parameters = num_parameters
super(PReLU, self).__init__()
self.weight = Parameter(torch.Tensor(num_parameters).fill_(init))
- def forward(self, input: Tensor) -> Tensor:
+ @weak_script_method
+ def forward(self, input):
return F.prelu(input, self.weight)
- def extra_repr(self) -> str:
+ def extra_repr(self):
return 'num_parameters={}'.format(self.num_parameters)