Layer Normalization
Layer normalization (LN) is a technique used in deep neural networks to stabilize training and improve performance by normalizing the activations of neurons within a layer. Current research focuses on understanding LN's geometric properties, its interaction with other normalization methods (like RMSNorm and Batch Normalization), and its impact on model stability and efficiency, particularly within transformer architectures and various applications such as natural language processing and image generation. These investigations aim to optimize LN's implementation, potentially leading to more efficient and robust deep learning models across diverse domains.
Papers
May 28, 2023
May 26, 2023
May 24, 2023
May 16, 2023
May 4, 2023
April 28, 2023
March 21, 2023
March 19, 2023
March 3, 2023
February 15, 2023
February 7, 2023
December 13, 2022
December 10, 2022
December 5, 2022
November 16, 2022
October 31, 2022
October 11, 2022
September 30, 2022
September 28, 2022