Skip to content

Instantly share code, notes, and snippets.

@goddoe
Created December 16, 2018 15:18
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save goddoe/dfd50258c0d6e996da265e0e94e77695 to your computer and use it in GitHub Desktop.
Save goddoe/dfd50258c0d6e996da265e0e94e77695 to your computer and use it in GitHub Desktop.
nomarlizations
Batch Norm:
(+) Stable if the batch size is large
(+) Robust (in train) to the scale & shift of input data
(+) Robust to the scale of weight vector
(+) Scale of update decreases while training
(-) Not good for online learning
(-) Not good for RNN, LSTM
(-) Different calculation between train and test
Weight Norm:
(+) Smaller calculation cost on CNN
(+) Well-considered about weight initialization
(+) Implementation is easy
(+) Robust to the scale of weight vector
(-) Compared with the others, might be unstable on training
(-) High dependence to input data
Layer Norm:
(+) Effective to small mini batch RNN
(+) Robust to the scale of input
(+) Robust to the scale and shift of weight matrix
(+) Scale of update decreases while training
(-) Might be not good for CNN (Batch Norm is better in some cases)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment