Tag

batch normalization

0 views collected around this technical thread.

IT Services Circle
IT Services Circle
May 2, 2025 · Artificial Intelligence

Understanding Gradient Vanishing in Deep Neural Networks and How to Mitigate It

The article explains why deep networks suffer from gradient vanishing—especially when using sigmoid or tanh activations—covers the underlying mathematics, compares activation functions, and presents practical techniques such as proper weight initialization, batch normalization, residual connections, and code examples to visualize the phenomenon.

Neural NetworksResNetactivation functions
0 likes · 7 min read
Understanding Gradient Vanishing in Deep Neural Networks and How to Mitigate It