Open Ended Text Generation
Open-ended text generation focuses on creating coherent and diverse text outputs from language models, aiming to improve the quality, creativity, and factual accuracy of generated content. Current research emphasizes developing advanced decoding strategies, such as contrastive search and adaptive sampling methods, to balance fluency with originality and mitigate issues like repetition and factual inaccuracies. These advancements are crucial for improving various applications, from personalized content generation to more reliable question-answering systems, while also addressing ethical concerns like bias and misinformation in generated text.
Papers
An Analysis of the Effects of Decoding Algorithms on Fairness in Open-Ended Language Generation
Jwala Dhamala, Varun Kumar, Rahul Gupta, Kai-Wei Chang, Aram Galstyan
Visualize Before You Write: Imagination-Guided Open-Ended Text Generation
Wanrong Zhu, An Yan, Yujie Lu, Wenda Xu, Xin Eric Wang, Miguel Eckstein, William Yang Wang