Sparsemax activation function.
tfa.activations.sparsemax(
logits: tfa.types.TensorLike
,
axis: int = -1
) -> tf.Tensor
For each batch , and class , compute sparsemax activation function:
See From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification.
Usage:
x = tf.constant([[-1.0, 0.0, 1.0], [-5.0, 1.0, 2.0]])
tfa.activations.sparsemax(x)
<tf.Tensor: shape=(2, 3), dtype=float32, numpy=
array([[0., 0., 1.],
[0., 0., 1.]], dtype=float32)>
Returns | |
---|---|
A Tensor , output of sparsemax transformation. Has the same type and
shape as logits .
|