Moral Preference Elicitation
Moral preference elicitation aims to quantitatively capture and represent human values for applications like aligning AI systems with human ethics. Current research focuses on improving the reliability and efficiency of elicitation methods, addressing challenges like the instability of moral judgments over time and the limitations of active learning algorithms in this context. These efforts are crucial for developing ethical AI, as accurately reflecting diverse human values is essential for building trustworthy and beneficial AI systems. Ongoing work explores various approaches, including text-based analysis of moral judgments and novel methods for synthesizing diverse value inputs into a coherent representation.
Papers
August 5, 2024
July 26, 2024
May 28, 2024