Automatic Metric
Automatic metrics aim to objectively assess the quality of outputs from natural language processing (NLP) and other AI systems, primarily by correlating with human judgments. Current research focuses on improving the accuracy and reliability of these metrics, particularly addressing their limitations in capturing nuanced aspects of quality, especially at higher performance levels, and developing metrics tailored to specific tasks (e.g., LaTeX formatting, video generation). This work is crucial for efficient and unbiased evaluation of NLP systems, enabling faster progress in model development and facilitating more rigorous comparisons across different approaches.
Papers
July 6, 2023
June 6, 2023
May 23, 2023
March 7, 2023
December 12, 2022
October 24, 2022
August 31, 2022
August 24, 2022
April 21, 2022
April 15, 2022
April 1, 2022