Moral Dilemma
Moral dilemmas, situations with conflicting ethical choices, are a central focus in evaluating the ethical implications of large language models (LLMs). Current research investigates how LLMs handle these dilemmas, analyzing their responses through various ethical frameworks and assessing biases related to gender, ethnicity, and cultural norms. This work is crucial for developing more responsible and equitable AI systems, impacting both the design of future models and the ethical guidelines governing their deployment in real-world applications. The goal is to move beyond simply measuring accuracy to understanding and mitigating the inherent biases and limitations of LLMs in navigating complex moral choices.
Papers
TRIAGE: Ethical Benchmarking of AI Models Through Mass Casualty Simulations
Nathalie Maria Kirch, Konstantin Hebenstreit, Matthias Samwald
Fine-Tuning Language Models for Ethical Ambiguity: A Comparative Study of Alignment with Human Responses
Pranav Senthilkumar, Visshwa Balasubramanian, Prisha Jain, Aneesa Maity, Jonathan Lu, Kevin Zhu