Latent Concept
Latent concept research focuses on uncovering the underlying representations and reasoning processes within large language models (LLMs) and other deep learning models, aiming to improve model interpretability and performance. Current research utilizes various techniques, including clustering algorithms to analyze attention head behavior and latent spaces, prompt engineering to elicit model knowledge, and causal modeling to understand the relationships between inputs, latent variables, and outputs. Understanding these latent concepts is crucial for enhancing model reliability, addressing biases, and improving the design of future AI systems across diverse applications, from question answering to image generation.
Papers
January 18, 2024
September 12, 2023
September 6, 2023
August 20, 2023
March 25, 2023
March 6, 2023
March 1, 2023
December 20, 2022
December 17, 2022
December 14, 2022
November 12, 2022
October 23, 2022
September 15, 2022
August 11, 2022
June 27, 2022
May 24, 2022
May 15, 2022
December 16, 2021