TensorFlow 2 version
|
View source on GitHub
|
Rectified Linear Unit.
tf.keras.activations.relu(
x, alpha=0.0, max_value=None, threshold=0
)
With default values, it returns element-wise max(x, 0).
Otherwise, it follows:
f(x) = max_value for x >= max_value,
f(x) = x for threshold <= x < max_value,
f(x) = alpha * (x - threshold) otherwise.
Arguments | |
|---|---|
x
|
A tensor or variable. |
alpha
|
A scalar, slope of negative section (default=0.).
|
max_value
|
float. Saturation threshold. |
threshold
|
float. Threshold value for thresholded activation. |
Returns | |
|---|---|
| A tensor. |
TensorFlow 2 version
View source on GitHub