Latent Concept
Latent concept research focuses on uncovering the underlying representations and reasoning processes within large language models (LLMs) and other deep learning models, aiming to improve model interpretability and performance. Current research utilizes various techniques, including clustering algorithms to analyze attention head behavior and latent spaces, prompt engineering to elicit model knowledge, and causal modeling to understand the relationships between inputs, latent variables, and outputs. Understanding these latent concepts is crucial for enhancing model reliability, addressing biases, and improving the design of future AI systems across diverse applications, from question answering to image generation.
Papers
November 1, 2024
October 30, 2024
October 7, 2024
September 5, 2024
July 20, 2024
July 18, 2024
June 26, 2024
June 18, 2024
June 8, 2024
June 4, 2024
May 23, 2024
May 12, 2024
April 19, 2024
April 18, 2024
April 16, 2024
March 18, 2024
January 18, 2024
September 12, 2023
September 6, 2023