GPT Model
Generative Pre-trained Transformer (GPT) models are large language models designed to generate human-like text, with research focusing on improving their accuracy, mitigating biases, and enhancing their applicability across diverse fields. Current research explores architectural improvements, such as optimizing attention mechanisms and employing sparsity techniques for efficient training, alongside investigations into bias mitigation strategies and the development of specialized GPT models for specific domains (e.g., biomedical text summarization, financial analysis). The impact of GPT models is significant, offering potential for automating tasks, improving accessibility to information, and advancing research in areas like natural language processing and scientific simulation, although concerns regarding bias and privacy remain active research areas.
Papers
SikuGPT: A Generative Pre-trained Model for Intelligent Information Processing of Ancient Texts from the Perspective of Digital Humanities
Liu Chang, Wang Dongbo, Zhao Zhixiao, Hu Die, Wu Mengcheng, Lin Litao, Shen Si, Li Bin, Liu Jiangfeng, Zhang Hai, Zhao Lianzheng
ArguGPT: evaluating, understanding and identifying argumentative essays generated by GPT models
Yikang Liu, Ziyin Zhang, Wanyang Zhang, Shisen Yue, Xiaojing Zhao, Xinyuan Cheng, Yiwen Zhang, Hai Hu