Auto Regressive Generation
Autoregressive generation focuses on creating sequential data, like text or images, by predicting each element based on previously generated ones. Current research emphasizes improving efficiency, particularly through parallel decoding methods like Jacobi decoding and techniques that reduce computational cost, such as those leveraging convolutional operators or dynamic resource allocation. These advancements are significant because they enable faster and more efficient generation of high-quality outputs across diverse applications, from text generation and image synthesis to real-time game simulation and time series forecasting. The resulting speed improvements are crucial for deploying these models in resource-constrained environments and real-time applications.
Papers
Accelerating Auto-regressive Text-to-Image Generation with Training-free Speculative Jacobi Decoding
Yao Teng, Han Shi, Xian Liu, Xuefei Ning, Guohao Dai, Yu Wang, Zhenguo Li, Xihui Liu
FutureFill: Fast Generation from Convolutional Sequence Models
Naman Agarwal, Xinyi Chen, Evan Dogariu, Vlad Feinberg, Daniel Suo, Peter Bartlett, Elad Hazan