# elu¶

paddle.fluid.layers.elu(x, alpha=1.0, name=None)[source]

ELU Activation Operator.

Applies the following element-wise computation on the input according to https://arxiv.org/abs/1511.07289.

$$out = \max(0, x) + \min(0, \alpha * (e^x - 1))$$

Parameters
• x (Variable) – The input is a multi-dimensional Tensor. The data type is float32 or float64

• alpha (FLOAT|1.0) – The alpha value of ELU

• name (str|None) – The default value is None. Normally there is no need for user to set this property. For more information, please refer to Name.

Returns

The output is a multi-dimensional Tensor which has same dimension and data type as the x

Return type

Variable

Examples

import paddle.fluid as fluid
import numpy as np

input_elu = np.array([[-1,6],[1,15.6]])
with fluid.dygraph.guard():
x = fluid.dygraph.to_variable(input_elu)
y = fluid.layers.elu(x, alpha=0.2)
print(y.numpy())
# [[-0.12642411  6.        ]
# [ 1.          15.6       ]]