Tag: vanishing gradient problem
-
The Vanishing Gradient Problem
The vanishing gradient problem can plague deep neural networks which consists of many hidden layers. It generally occurs when the derivatives of the activation functions are less than one. This leads to the multiplication of small numbers during back-propagation, which in turn shrink the gradients exponentially. As a result of this issue, the network will…