Better Zero
"Better Zero" research focuses on improving the performance of machine learning models in zero-shot and few-shot learning scenarios, minimizing the need for large, labeled training datasets. Current efforts concentrate on developing novel prompt engineering techniques, leveraging pre-trained large language models (LLMs) and vision-language models (VLMs), and designing efficient algorithms for proxy search and model adaptation. This research is significant because it addresses the limitations of data-hungry models, potentially enabling wider application of AI in resource-constrained domains and accelerating the development of more generalizable AI systems.
Papers
October 15, 2024
October 7, 2024
September 30, 2024
September 26, 2024
August 12, 2024
June 1, 2024
May 30, 2024
May 6, 2024
April 30, 2024
April 17, 2024
March 26, 2024
March 4, 2024
February 19, 2024
February 15, 2024
January 8, 2024
December 13, 2023
December 4, 2023
November 4, 2023
October 6, 2023