Rate this Page
โ˜… โ˜… โ˜… โ˜… โ˜…

torch.nn.functional.gelu#

torch.nn.functional.gelu(input, approximate='none') โ†’ Tensor#

When the approximate argument is โ€˜noneโ€™, it applies element-wise the function GELU(x)=xโˆ—ฮฆ(x)\text{GELU}(x) = x * \Phi(x)

where ฮฆ(x)\Phi(x) is the Cumulative Distribution Function for Gaussian Distribution.

When the approximate argument is โ€˜tanhโ€™, Gelu is estimated with

GELU(x)=0.5โˆ—xโˆ—(1+Tanh(2/ฯ€โˆ—(x+0.044715โˆ—x3)))\text{GELU}(x) = 0.5 * x * (1 + \text{Tanh}(\sqrt{2 / \pi} * (x + 0.044715 * x^3)))

See Gaussian Error Linear Units (GELUs).