Parameter Sharing
Parameter sharing in machine learning aims to reduce computational costs and improve efficiency by reusing model parameters across different tasks, layers, or agents. Current research focuses on developing novel parameter-sharing strategies within various architectures, including transformers, convolutional neural networks, and multi-agent reinforcement learning models, often incorporating techniques like mixture-of-experts and adaptive gating mechanisms to optimize resource allocation and performance. This research is significant because efficient parameter sharing is crucial for deploying large-scale models in resource-constrained environments and for improving the scalability and training efficiency of complex systems.
Papers
November 14, 2024
October 28, 2024
October 11, 2024
October 1, 2024
August 14, 2024
July 20, 2024
June 22, 2024
June 21, 2024
June 5, 2024
May 29, 2024
May 7, 2024
April 18, 2024
March 30, 2024
March 28, 2024
March 5, 2024
February 13, 2024
January 20, 2024
December 26, 2023
December 14, 2023