Generative Transformer

Generative transformers are neural network architectures designed to generate various data types, including text, images, and time series, by learning complex patterns and relationships within input data. Current research focuses on improving efficiency (e.g., through analog computing and non-autoregressive methods), enhancing controllability (e.g., via multimodal prompting and structure-guided generation), and mitigating limitations like repetitive outputs and spurious correlations. These advancements are significantly impacting fields like natural language processing, computer vision, and time series forecasting, enabling applications ranging from improved medical report generation to more efficient video editing and advanced AI-driven design tools.

Papers