Gating Mechanism
Gating mechanisms are control structures in machine learning models that selectively activate or suppress information flow, improving efficiency and performance. Current research focuses on developing adaptive gating strategies within various architectures, including Mixture-of-Experts (MoE) models, transformers, and graph neural networks, often employing techniques like stochastic filtering, sensitivity-based adjustments, and learnable prompts to optimize resource allocation and enhance model robustness. These advancements are significant because they lead to more efficient and accurate models across diverse applications, from natural language processing and computer vision to time-series prediction and resource-constrained environments.
Papers
Sigmoid Gating is More Sample Efficient than Softmax Gating in Mixture of Experts
Huy Nguyen, Nhat Ho, Alessandro Rinaldo
DEGAP: Dual Event-Guided Adaptive Prefixes for Templated-Based Event Argument Extraction with Slot Querying
Guanghui Wang, Dexi Liu, Jian-Yun Nie, Qizhi Wan, Rong Hu, Xiping Liu, Wanlong Liu, Jiaming Liu