Capturing Extreme Activation
Capturing extreme activations in neural networks aims to improve model efficiency, robustness, and controllability. Current research focuses on developing novel activation functions and dynamic activation techniques, applied to various architectures including CNNs, RNNs, LSTMs, and large language models, to enhance performance in tasks like intrusion detection and out-of-distribution detection. These advancements are significant because they offer ways to optimize model inference speed, mitigate overconfidence in predictions, and enable more effective control over model behavior without extensive retraining, leading to more efficient and reliable AI systems.
Papers
August 21, 2024
May 30, 2024
May 21, 2024
February 14, 2024
November 10, 2023
August 20, 2023