Moral Evaluation
Moral evaluation in artificial intelligence focuses on assessing the ethical reasoning and decision-making capabilities of AI systems, particularly large language models (LLMs), and ensuring alignment with human moral standards. Current research employs benchmarks and modified Turing tests to evaluate LLMs' performance across various moral dilemmas, analyzing consistency in abstract and concrete moral judgments and investigating potential biases. This work is crucial for mitigating the risks of harmful AI-driven decisions and promoting the responsible development and deployment of AI systems in diverse societal contexts.
Papers
June 6, 2024
May 17, 2024
April 3, 2024
November 16, 2023
November 14, 2023
May 19, 2023