Translations:Batch Normalization Accelerating Deep Network Training/23/en: Difference between revisions
(Importing a new version from external source) |
(Importing a new version from external source) |
||
| Line 1: | Line 1: | ||
{{Term|batch normalization}} became one of the most ubiquitous components in {{Term|deep learning}} architectures. It was adopted almost universally in convolutional networks throughout the late 2010s and remains standard in many architectures. The technique's success inspired a family of normalization methods, including {{Term|layer normalization}} (preferred in {{Term|transformer|Transformers}} and recurrent networks), instance normalization (used in style transfer), and group normalization (useful for small batch sizes). | |||
Latest revision as of 21:40, 27 April 2026
batch normalization became one of the most ubiquitous components in deep learning architectures. It was adopted almost universally in convolutional networks throughout the late 2010s and remains standard in many architectures. The technique's success inspired a family of normalization methods, including layer normalization (preferred in Transformers and recurrent networks), instance normalization (used in style transfer), and group normalization (useful for small batch sizes).