Attribution Model
Attribution models aim to explain the internal workings of complex models, particularly deep learning systems, by identifying which input features most influence their outputs. Current research focuses on improving the accuracy and efficiency of these models, addressing challenges like handling diverse input formats (e.g., text, images), mitigating biases, and ensuring robustness against adversarial attacks. This work is crucial for enhancing the transparency and trustworthiness of AI systems, enabling better debugging, fairer decision-making, and improved understanding of model behavior across various applications, including natural language processing and network fault diagnosis.
Papers
July 31, 2024
May 2, 2024
April 16, 2024
February 5, 2024
September 19, 2023
February 23, 2022
February 10, 2022