TF 2.0 is out! Get hands-on practice at TF World, Oct 28-31. Use code TF20 for 20% off select passes. Register now

tf.nn.leaky_relu

TensorFlow 1 version View source on GitHub

Compute the Leaky ReLU activation function.

Aliases:

  • tf.compat.v1.nn.leaky_relu
  • tf.compat.v2.nn.leaky_relu
tf.nn.leaky_relu(
    features,
    alpha=0.2,
    name=None
)

Source: Rectifier Nonlinearities Improve Neural Network Acoustic Models. AL Maas, AY Hannun, AY Ng - Proc. ICML, 2013.

Args:

  • features: A Tensor representing preactivation values. Must be one of the following types: float16, float32, float64, int32, int64.
  • alpha: Slope of the activation function at x < 0.
  • name: A name for the operation (optional).

Returns:

The activation value.