Protein Language Model
Protein language models (PLMs) leverage the power of natural language processing techniques to analyze and generate protein sequences, aiming to improve our understanding of protein structure, function, and evolution. Current research focuses on enhancing PLMs through techniques like instruction tuning, incorporating structural information (e.g., using graph neural networks or contact maps), and employing various training strategies such as contrastive learning and reinforcement learning to optimize for specific properties or tasks. These advancements are significantly impacting various fields, including drug discovery, protein engineering, and the broader understanding of biological processes by enabling more accurate predictions and efficient design of proteins with desired characteristics.
Papers
Metalic: Meta-Learning In-Context with Protein Language Models
Jacob Beck, Shikha Surana, Manus McAuliffe, Oliver Bent, Thomas D. Barrett, Juan Jose Garau Luis, Paul Duckworth
Steering Masked Discrete Diffusion Models via Discrete Denoising Posterior Prediction
Jarrid Rector-Brooks, Mohsin Hasan, Zhangzhi Peng, Zachary Quinn, Chenghao Liu, Sarthak Mittal, Nouha Dziri, Michael Bronstein, Yoshua Bengio, Pranam Chatterjee, Alexander Tong, Avishek Joey Bose
Conditional Enzyme Generation Using Protein Language Models with Adapters
Jason Yang, Aadyot Bhatnagar, Jeffrey A. Ruffolo, Ali Madani
Structure-Enhanced Protein Instruction Tuning: Towards General-Purpose Protein Understanding
Wei Wu, Chao Wang, Liyi Chen, Mingze Yin, Yiheng Zhu, Kun Fu, Jieping Ye, Hui Xiong, Zheng Wang