Data Skeptic

[MINI] The Vanishing Gradient

15:16 | Jun 30th, 2017

This episode discusses the vanishing gradient - a problem that arises when training deep neural networks in which nearly all the gradients are very close to zero by the time back-propagation has reached the first hidden layer. This makes learning vir...Show More