Hyper Representation
Hyper-representation research focuses on learning compact, task-agnostic representations of neural network weights to understand and manipulate model behavior. Current work utilizes autoencoders and other methods to create these representations, enabling tasks like generating new network weights with specific properties, improving transfer learning, and facilitating efficient fine-tuning across diverse architectures and tasks. This approach promises to enhance model interpretability, improve training efficiency, and potentially lead to more adaptable and powerful foundation models for various applications.
Papers
October 7, 2024
June 14, 2024
September 29, 2022
July 26, 2022
July 22, 2022
June 6, 2022
March 8, 2022