Editing Benchmark
Editing benchmarks evaluate methods for modifying large language models (LLMs) and other models, such as recommendation systems and vision-language models, without full retraining. Current research focuses on developing benchmarks for various editing tasks, including correcting factual errors, updating knowledge based on events, and modifying image generation outputs, often using natural language instructions to specify the desired changes. These benchmarks are crucial for evaluating the effectiveness of different editing techniques and driving progress in areas like improving model reliability, mitigating biases, and enhancing user experience in applications ranging from personalized recommendations to image editing software.
Papers
June 6, 2024
March 12, 2024
February 20, 2024
December 22, 2023
November 27, 2023
October 2, 2023