Agreement Metric
Agreement metrics quantify the consistency of predictions or annotations across different models, annotators, or time points, aiming to improve the reliability and trustworthiness of AI systems and analyses. Current research focuses on developing novel metrics tailored to specific tasks (e.g., sequence annotation, medical image classification, multi-party conversations), often incorporating transformer networks or graph-based methods to handle complex data structures and relationships. These advancements are crucial for enhancing the explainability and robustness of AI models, particularly in high-stakes applications like healthcare and human-robot interaction, where reliable agreement is paramount.
Papers
Detecting Agreement in Multi-party Conversational AI
Laura Schauer, Jason Sweeney, Charlie Lyttle, Zein Said, Aron Szeles, Cale Clark, Katie McAskill, Xander Wickham, Tom Byars, Daniel Hernández Garcia, Nancie Gunson, Angus Addlesee, Oliver Lemon
Detecting agreement in multi-party dialogue: evaluating speaker diarisation versus a procedural baseline to enhance user engagement
Angus Addlesee, Daniel Denley, Andy Edmondson, Nancie Gunson, Daniel Hernández Garcia, Alexandre Kha, Oliver Lemon, James Ndubuisi, Neil O'Reilly, Lia Perochaud, Raphaël Valeri, Miebaka Worika