QueST
"QueST," a term appearing across diverse research areas, broadly refers to the pursuit of improved efficiency and interpretability in various machine learning contexts. Current research focuses on developing novel algorithms and architectures, such as graph transformers and quantized skill transformers, to enhance model performance in tasks ranging from long-context language modeling to robot control and quantum circuit reliability estimation. These efforts aim to address limitations in existing methods, particularly concerning generalization, scalability, and the explainability of model decisions, ultimately impacting fields from robotics and natural language processing to quantum computing.
Papers
September 30, 2024
September 19, 2024
August 2, 2024
July 22, 2024
June 16, 2024
May 30, 2024
May 28, 2024
February 6, 2024
November 8, 2023
September 25, 2023
August 31, 2023
August 3, 2023
July 8, 2023
May 19, 2023
October 30, 2022
February 26, 2022