Module juice::layers::activation::tanh
source · Expand description
Applies the nonlinear TanH function.
Non-linearity activation function: y = sinh(x) / cosh(x)
You might consider using ReLU as an alternative.
ReLU, compared to TanH:
- reduces the likelyhood of vanishing gradients
- increases the likelyhood of a more beneficial sparse representation
- can be computed faster
- is therefore the most popular activation function in DNNs as of this writing (2016).
Structs§
- TanH Activation Layer