Activation Functions
Published on: September 20, 2021
Table of Content
Sigmoid Function
Function:
Derivative:
Tanh Function
Function:
Derivative:
Rectified Linear Unit (ReLU)
Function:
Derivative:
Leaky ReLU
Function:
Derivative:
Parametric ReLU
Function:
Derivative:
Exponential Linear Unit (ELU)
Function:
Derivative:
Scaled Exponential Linear Unit (SELU)
Function:
Derivative:
Gaussian Error Linear Unit (GELU)
Function:
Derivative:
Sigmoid-Weighted Linear Unit (SiLU) / Swish
Function:
Derivative:
Softmax Function
Function:
Softplus
Function:
Derivative:
Mish
Function:
Derivative:
Resources
- https://en.wikipedia.org/wiki/Activation_function
- https://medium.com/the-theory-of-everything/understanding-activation-functions-in-neural-networks-9491262884e0
- https://www.v7labs.com/blog/neural-networks-activation-functions
- https://mlfromscratch.com/activation-functions-explained/#/
- https://ml-cheatsheet.readthedocs.io/en/latest/activation_functions.html