def _activate()

in crlapi/sl/architectures/firefly_vgg/sp/module.py [0:0]


    def _activate(self, x):
        if self.actv_fn == 'relu':
            return F.relu(x)
        elif self.actv_fn == 'leaky_relu':
            return F.leaky_relu(x, self.leaky_alpha)
        elif self.actv_fn == 'swish':
            return x * torch.sigmoid(x)
        elif self.actv_fn == 'rbf':
            return (-x.pow(2)/2).exp()
        elif self.actv_fn == 'sigmoid':
            return torch.sigmoid(x)
        elif self.actv_fn == 'tanh':
            return torch.tanh(x)
        elif self.actv_fn == 'softplus':
            return F.softplus(x)
        elif self.actv_fn == 'none':
            return x
        else:
            raise NotImplementedError