User Tools

Site Tools


ml:normalization

Normalization

Normalization can improve the optimizer's ability to train a neural network. There are two main categories of normalization procedures: activation normalization and weight normalization (Shen 2020).

Overviews

Activation Normalization Schemes

Batch Normalization

Layer Normalization

Weight Normalization Schemes

Weight Normalization

  • Weight normalization is billed as an alternative to batch normalization. Salimans & Kingma 2016 - Weight Normalization: A Simple Reparameterization to Accelerate Training of Deep Neural Networks “…improve the conditioning of the optimization problem and we speed up convergence of stochastic gradient descent. Our reparameterization is inspired by batch normalization but does not introduce any dependencies between the examples in a minibatch. This means that our method can also be applied successfully to recurrent models such as LSTMs and to noise-sensitive applications such as deep reinforcement learning or generative models, for which batch normalization is less well suited.” See section 2.2 Relation to batch normalization.

Other or Uncategorized Schemes

ml/normalization.txt · Last modified: 2023/06/15 07:36 by 127.0.0.1

Donate Powered by PHP Valid HTML5 Valid CSS Driven by DokuWiki