Transformer Based LLM

Transformer-based large language models (LLMs) are a class of deep learning models designed to process and generate human-like text, with applications ranging from cybersecurity to materials science. Current research focuses on improving efficiency through techniques like low-rank parameterization and binarization, as well as enhancing capabilities in long-context processing and mitigating privacy risks via improved membership inference attack defenses. These advancements aim to reduce computational costs while maintaining or improving performance across various tasks, impacting fields requiring natural language understanding and generation.

Papers