Interpretable Layer
Interpretable layers in neural networks aim to enhance the transparency and understandability of deep learning models, addressing the "black box" problem. Current research focuses on developing novel layer architectures, such as simplicial maps and piecewise linear functions, and integrating them into existing models like CNNs and transformers, often using techniques like attention mechanisms and class activation mapping to highlight relevant features. This work is significant because it improves trust and allows for better debugging and refinement of complex models, particularly in high-stakes applications like legal decision-making and medical diagnosis where understanding model reasoning is crucial.
Papers
October 23, 2024
October 1, 2024
March 22, 2024
February 19, 2024
February 12, 2024
February 1, 2024
October 5, 2023
March 28, 2023
March 8, 2022
January 1, 2022