brelu¶
- paddle.fluid.layers.nn. brelu ( x, t_min=0.0, t_max=24.0, name=None ) [source]
- 
         BRelu Activation Operator. \(out = \min(\max(x, t_{min}), t_{max})\) - Parameters
- 
           - x (Variable) – The input is a multi-dimensional Tensor. The data type is float32, float64 
- t_min (FLOAT|0.0) – The min marginal value of BRelu 
- t_max (FLOAT|24.0) – The max marginal value of BRelu 
- name (str|None) – The default value is None. Normally there is no need for user to set this property. For more information, please refer to Name. 
 
- Returns
- 
           The output is a multi-dimensional Tensor which has same dimension and data type as the X
- Return type
- 
           Variable 
 Examples: import paddle.fluid as fluid import paddle import numpy as np paddle.enable_static() input_brelu = np.array([[-1,6],[1,15.6]]) with fluid.dygraph.guard(): x = fluid.dygraph.to_variable(input_brelu) y = fluid.layers.brelu(x, t_min=1.0, t_max=10.0) print(y.numpy()) #[[ 1. 6.] #[ 1. 10.]] 
