Natural Language Generation
Natural Language Generation (NLG) focuses on creating human-readable text from structured data or other inputs. Current research emphasizes improving the accuracy and fluency of generated text, particularly addressing issues like information omission and biases stemming from the dominance of English in training data, and exploring diverse model architectures such as transformers. A significant focus is on developing more reliable and nuanced evaluation methods, moving beyond simple metrics to incorporate human judgment and address challenges like detecting hallucinations and ensuring fairness. These advancements have implications for various applications, including improved search engine advertising, more effective educational tools, and enhanced accessibility for low-resource languages.
Papers
IndicNLG Benchmark: Multilingual Datasets for Diverse NLG Tasks in Indic Languages
Aman Kumar, Himani Shrotriya, Prachi Sahu, Raj Dabre, Ratish Puduppully, Anoop Kunchukuttan, Amogh Mishra, Mitesh M. Khapra, Pratyush Kumar
Faithfulness in Natural Language Generation: A Systematic Survey of Analysis, Evaluation and Optimization Methods
Wei Li, Wenhao Wu, Moye Chen, Jiachen Liu, Xinyan Xiao, Hua Wu