ReLU
- ReLU = Rectified Linear Unit
- It solves the Vanishing Gradient problem
- But introduces Dying ReLU problem
- Most popular function for CNN or even #deep-learning models
[!def] ReLU Formula
$$
f(x) = max(0, x)
$$
Derivative of ReLU
$$
f'(x) = 0 \text{ if x <= 0}
$$
$$
f'(x) = 1 \text{ if x > 0}
$$