Classes
| Activation<T extends TNumber> | Abstract base class for Activations
Note: The |
| ELU<T extends TFloating> | Exponential linear unit. |
| Exponential<T extends TFloating> | Exponential activation function. |
| HardSigmoid<T extends TFloating> | Hard sigmoid activation. |
| Linear<U extends TNumber> | Linear activation function (pass-through). |
| ReLU<T extends TNumber> | Rectified Linear Unit(ReLU) activation. |
| SELU<T extends TFloating> | Scaled Exponential Linear Unit (SELU). |
| Sigmoid<T extends TFloating> | Sigmoid activation. |
| Softmax<T extends TFloating> | Softmax converts a real vector to a vector of categorical probabilities. |
| Softplus<T extends TFloating> | Softplus activation function, softplus(x) = log(exp(x) + 1). |
| Softsign<T extends TFloating> | Softsign activation function, softsign(x) = x / (abs(x) + 1). |
| Swish<T extends TFloating> | Swish activation function. |
| Tanh<T extends TFloating> | Hyperbolic tangent activation function. |