Zeroth Order Optimization
Zeroth-order optimization (ZO) tackles the challenge of optimizing functions where gradients are unavailable or computationally expensive, a common hurdle in complex machine learning problems. Current research focuses on improving the efficiency and scalability of ZO methods, particularly for large language model (LLM) fine-tuning and federated learning, employing techniques like randomized gradient estimation and sparse parameter updates within algorithms such as ZO-SGD and MeZO. These advancements are significant because they enable memory-efficient training of large models on resource-constrained devices and facilitate privacy-preserving collaborative learning, impacting diverse fields from drug discovery to reinforcement learning.
Papers
October 13, 2024
October 9, 2024
October 3, 2024
September 24, 2024
September 23, 2024
August 17, 2024
June 20, 2024
February 24, 2024
February 18, 2024
February 8, 2024
October 3, 2023
September 26, 2023
May 27, 2023
March 7, 2023
January 27, 2023
October 27, 2022
June 23, 2022