Contrastive Search
Contrastive search is a decoding method for language models aiming to generate higher-quality, more diverse, and coherent text by contrasting generated sequences against each other or against a source text. Current research focuses on refining contrastive search algorithms, such as incorporating uncertainty estimation or adaptive penalties, and applying them to diverse tasks including text generation, image analysis (e.g., lymph node detection), and neural architecture search. These improvements address limitations of previous methods, like generating repetitive or hallucinated text, and demonstrate significant performance gains across various benchmarks, highlighting the importance of contrastive approaches in improving the capabilities of large language models and other AI systems.
Papers
Cluster-norm for Unsupervised Probing of Knowledge
Walter Laurito, Sharan Maiya, Grégoire Dhimoïla, Owen, Yeung, Kaarel Hänni
Adaptive Contrastive Search: Uncertainty-Guided Decoding for Open-Ended Text Generation
Esteban Garces Arias, Julian Rodemann, Meimingwei Li, Christian Heumann, Matthias Aßenmacher