elu(x, alpha=1.0, name=None)
ELU Activation Operator.
Applies the following element-wise computation on the input according to https://arxiv.org/abs/1511.07289.
$$out = max(0, x) + min(0, alpha * (e^x - 1))$$
x (Variable) – The input is a multi-dimensional Tensor. The data type is float32 or float64
alpha (FLOAT|1.0) – The alpha value of ELU
name (str|None) – The default value is None. Normally there is no need for user to set this property. For more information, please refer to Name.
The output is a multi-dimensional Tensor which has same dimension and data type as the
- Return type
import paddle.fluid as fluid import numpy as np input_elu = np.array([[-1,6],[1,15.6]]) with fluid.dygraph.guard(): x = fluid.dygraph.to_variable(input_elu) y = fluid.layers.elu(x, alpha=0.2) print(y.numpy()) # [[-0.12642411 6. ] # [ 1. 15.6 ]]