Human Value

Research on human values in AI focuses on aligning artificial intelligence systems with diverse human values and preferences, aiming to mitigate potential harms and ensure beneficial use. Current efforts concentrate on developing methods to represent and reason about these values within AI models, often employing large language models (LLMs) and reinforcement learning techniques, and evaluating alignment through various benchmarks and metrics. This research is crucial for building trustworthy and ethical AI systems, impacting both the development of safer AI technologies and the broader understanding of human values themselves. The field is actively exploring how to address the complexities of value pluralism and the challenges of accurately capturing and representing the nuances of human ethical decision-making.

Papers