Automatic Metric
Automatic metrics aim to objectively assess the quality of outputs from natural language processing (NLP) and other AI systems, primarily by correlating with human judgments. Current research focuses on improving the accuracy and reliability of these metrics, particularly addressing their limitations in capturing nuanced aspects of quality, especially at higher performance levels, and developing metrics tailored to specific tasks (e.g., LaTeX formatting, video generation). This work is crucial for efficient and unbiased evaluation of NLP systems, enabling faster progress in model development and facilitating more rigorous comparisons across different approaches.
Papers
November 13, 2024
October 15, 2024
October 10, 2024
October 8, 2024
September 15, 2024
September 10, 2024
August 17, 2024
July 29, 2024
July 3, 2024
June 21, 2024
June 6, 2024
May 28, 2024
April 17, 2024
February 15, 2024
January 29, 2024
January 2, 2024
November 11, 2023
November 3, 2023
September 19, 2023