AI Blog

AI Blog

by Michele Laurelli

Weight Initialization

/weɪt ɪˌnɪʃəlaɪˈzeɪʃən/
Technique
Definition

Methods for setting initial values of neural network weights before training begins.

Proper initialization is crucial for effective training. Random initialization breaks symmetry. Methods like Xavier/Glorot (for tanh/sigmoid) and He initialization (for ReLU) ensure gradients flow well initially.

Examples

1

Xavier initialization for tanh layers

2

He initialization for ReLU layers

3

Zero initialization for biases

Michele Laurelli - AI Research & Engineering