Translations:Batch Normalization/3/en

    From Marovi AI

    The original motivation for batch normalization was to address internal covariate shift — the phenomenon where the distribution of each layer's inputs changes during training as the parameters of preceding layers are updated. This shifting distribution forces each layer to continuously adapt, slowing down convergence and requiring careful initialization and small learning rates.