Risk Taxonomy
Risk taxonomies are structured classifications of potential hazards associated with emerging technologies, particularly large language models (LLMs) and other AI systems. Current research focuses on developing comprehensive taxonomies that encompass various risk categories, from biases and safety violations to security vulnerabilities and ethical concerns, often employing natural language processing (NLP) techniques like topic modeling to analyze large datasets of user interactions and incident reports. These taxonomies are crucial for benchmarking model safety, informing the development of mitigation strategies, and ultimately promoting the responsible design and deployment of AI systems across diverse applications.
Papers
AIR-Bench 2024: A Safety Benchmark Based on Risk Categories from Regulations and Policies
Yi Zeng, Yu Yang, Andy Zhou, Jeffrey Ziwei Tan, Yuheng Tu, Yifan Mai, Kevin Klyman, Minzhou Pan, Ruoxi Jia, Dawn Song, Percy Liang, Bo Li
Unveiling Disparities in Maternity Care: A Topic Modelling Approach to Analysing Maternity Incident Investigation Reports
Georgina Cosma, Mohit Kumar Singh, Patrick Waterson, Gyuchan Thomas Jun, Jonathan Back