What is Weight Initialization for Neural Networks?
What's this blog post about?
Weight Initialization plays a significant role in deep feedforward neural networks' training process. Xavier Glorot and Yoshua Bengio highlighted the issue of using normal distribution for initializing weights with mean 0 and variance 1, which contributes to unstable gradients. To tackle these problems, new techniques have been introduced. This video discusses these methods, their differences, and ideal activation functions they correspond to.
Company
AssemblyAI
Date published
Jan. 31, 2022
Author(s)
Misra Turp
Word count
85
Hacker News points
None found.
Language
English