Justification Theory

Justification theory seeks to provide rigorous, understandable explanations for the outputs of complex systems, particularly in AI applications. Current research focuses on developing methods to generate these justifications, employing techniques like multi-agent systems with large language models and Bayesian approaches for extracting explanations from "black box" deep neural networks. This work is crucial for building trust and transparency in AI systems across diverse fields, from healthcare and law to recommender systems, by ensuring that decisions are not only accurate but also readily interpretable and justifiable. The ultimate goal is to create AI that is not only effective but also accountable and explainable.

Papers