ez_transfer.layers¶
activations¶
-
easytransfer.layers.activations.
gelu
(x)[source]¶ Gaussian Error Linear Unit. Original Implementation of the gelu activation function in Google Bert repo when initially created.
For information: OpenAI GPT's gelu is slightly different (and gives slightly different results): 0.5 * x * (1 + torch.tanh(math.sqrt(2 / math.pi) * (x + 0.044715 * torch.pow(x, 3)))) Also see https://arxiv.org/abs/1606.08415
-
easytransfer.layers.activations.
gelu_new
(x)[source]¶ Gaussian Error Linear Unit. This is a smoother version of the RELU. Original paper: https://arxiv.org/abs/1606.08415 :param x: float Tensor to perform activation.
Returns: x with the GELU activation applied.