Code Generation Model
Code generation models leverage large language models (LLMs) to automatically produce source code from natural language descriptions or other inputs, aiming to boost developer productivity and automate programming tasks. Current research emphasizes improving code quality and robustness, including developing more efficient prompt engineering techniques and addressing issues like security vulnerabilities, bias, and the generation of inefficient or hallucinated code. These advancements are significant for both the software engineering community, offering tools to enhance coding efficiency, and the broader AI field, providing a rich testbed for evaluating and improving LLMs' capabilities in complex, structured data generation.
Papers
A Static Evaluation of Code Completion by Large Language Models
Hantian Ding, Varun Kumar, Yuchen Tian, Zijian Wang, Rob Kwiatkowski, Xiaopeng Li, Murali Krishna Ramanathan, Baishakhi Ray, Parminder Bhatia, Sudipta Sengupta, Dan Roth, Bing Xiang
Computing Education in the Era of Generative AI
Paul Denny, James Prather, Brett A. Becker, James Finnie-Ansley, Arto Hellas, Juho Leinonen, Andrew Luxton-Reilly, Brent N. Reeves, Eddie Antonio Santos, Sami Sarsa
Uncovering and Quantifying Social Biases in Code Generation
Yan Liu, Xiaokang Chen, Yan Gao, Zhe Su, Fengji Zhang, Daoguang Zan, Jian-Guang Lou, Pin-Yu Chen, Tsung-Yi Ho
ALGO: Synthesizing Algorithmic Programs with LLM-Generated Oracle Verifiers
Kexun Zhang, Danqing Wang, Jingtao Xia, William Yang Wang, Lei Li
Generation-Augmented Query Expansion For Code Retrieval
Dong Li, Yelong Shen, Ruoming Jin, Yi Mao, Kuan Wang, Weizhu Chen
ReCode: Robustness Evaluation of Code Generation Models
Shiqi Wang, Zheng Li, Haifeng Qian, Chenghao Yang, Zijian Wang, Mingyue Shang, Varun Kumar, Samson Tan, Baishakhi Ray, Parminder Bhatia, Ramesh Nallapati, Murali Krishna Ramanathan, Dan Roth, Bing Xiang