Probabilistic Transformer
Probabilistic transformers are emerging as powerful tools for modeling uncertainty and complex dependencies in various data types, aiming to improve the robustness and interpretability of traditional transformer models. Current research focuses on adapting transformer architectures to incorporate probabilistic representations, often using Bayesian methods or conditional generative models, for tasks ranging from 3D shape generation and speech recognition to time series forecasting and outlier detection. This approach offers significant potential for advancing fields like machine learning, natural language processing, and scientific modeling by providing more reliable and explainable predictions in the face of noisy or ambiguous data.
Papers
Probabilistic Transformer: Modelling Ambiguities and Distributions for RNA Folding and Molecule Design
Jörg K. H. Franke, Frederic Runge, Frank Hutter
Sample-Efficient Optimisation with Probabilistic Transformer Surrogates
Alexandre Maraval, Matthieu Zimmer, Antoine Grosnit, Rasul Tutunov, Jun Wang, Haitham Bou Ammar