CodeT5+: Open Code Large Language Models for Code Understanding and Generation

被引:0
作者
Wang, Yue [1 ]
Le, Hung [1 ]
Gotmare, Akhilesh Deepak [1 ]
Bui, Nghi D. Q. [1 ]
Li, Junnan [1 ]
Hoi, Steven C. H. [1 ]
机构
[1] Salesforce AI Res, San Francisco, CA 94105 USA
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) pretrained on vast source code have achieved prominent progress in code intelligence. However, existing code LLMs have two main limitations. First, they often adopt a specific architecture (encoder-only or decoder-only) or rely on a unified encoder-decoder network for different downstream tasks, lacking the flexibility to operate in the optimal architecture for a specific task. Secondly, they often employ a limited set of pretraining objectives which might not be relevant to some tasks and hence result in substantial performance degrade. To address these limitations, we propose "CodeT5+", a family of encoder-decoder LLMs for code in which component modules can be flexibly combined to suit a wide range of code tasks. Such flexibility is enabled by our proposed mixture of pretraining objectives, which cover span denoising, contrastive learning, text-code matching, and causal LM pretraining tasks, on both unimodal and bimodal multilingual code corpora. Furthermore, we propose to initialize CodeT5+ with frozen off-the-shelf LLMs without training from scratch to efficiently scale up our models, and explore instruction-tuning to align with natural language instructions. We extensively evaluate CodeT5+ on over 20 code-related benchmarks in different settings, including zero-shot, finetuning, and instruction-tuning. We observe state-of-the-art (SoTA) performance on various code-related tasks, and our instruction-tuned CodeT5+ 16B achieves new SoTA results of 35.0% pass@1 and 54.5% pass@10 on the HumanEval code generation task against other open code LLMs, even surpassing the OpenAI code-cushman-001 model.
引用
收藏
页码:1069 / 1088
页数:20
相关论文
共 50 条
[21]   Anchor Attention, Small Cache: Code Generation With Large Language Models [J].
Zhang, Xiangyu ;
Zhou, Yu ;
Yang, Guang ;
Gall, Harald C. ;
Chen, Taolue .
IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2025, 51 (06) :1866-1881
[22]   Private-library-oriented code generation with large language models [J].
Zan, Daoguang ;
Chen, Bei ;
Gong, Yongshun ;
Cao, Junzhi ;
Zhang, Fengji ;
Wu, Bingchao ;
Guan, Bei ;
Yin, Yilong ;
Wang, Yongji .
KNOWLEDGE-BASED SYSTEMS, 2025, 326
[23]   Multi-stage guided code generation for Large Language Models [J].
Han, Yewei ;
Lyu, Chen .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
[24]   Is Your Code Generated by ChatGPT Really Correct? Rigorous Evaluation of Large Language Models for Code Generation [J].
Liu, Jiawei ;
Xia, Chunqiu Steven ;
Wang, Yuyao ;
Zhang, Lingming .
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
[25]   Large Language Models of Code Fail at Completing Code with Potential Bugs [J].
Tuan Dinh ;
Zhao, Jinman ;
Tan, Samson ;
Negrinho, Renato ;
Lausen, Leonard ;
Zha, Sheng ;
Karypis, George .
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
[26]   Parallel Code Generation with Large Language Model [J].
Huang, Xiaowen ;
Zhang, Xu ;
Tao, Lvfang ;
Mao, Renjie ;
Zhu, Wenxi ;
Deng, Mingwen ;
Zhou, Nan ;
Meng, Jintao ;
Wei, Yanjie ;
Zhou, Amelie Chi ;
Wang, Bingqiang ;
Feng, Shengzhong .
PROCEEDINGS OF THE 1ST FASTCODE PROGRAMMING CHALLENGE, FCPC 2025, 2025, :1-7
[27]   Invited Paper: VerilogEval: Evaluating Large Language Models for Verilog Code Generation [J].
Liu, Mingjie ;
Pinckney, Nathaniel ;
Khailany, Brucek ;
Ren, Haoxing .
2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
[28]   Code-level quantum circuit generation based on large language models [J].
He, Zhimin ;
Li, Guohong ;
Situ, Haozhen ;
Zhou, Yan ;
Zheng, Shenggen ;
Li, Lvzhou .
SCIENTIA SINICA-PHYSICA MECHANICA & ASTRONOMICA, 2025, 55 (04)
[29]   FormalEval: A Method for Automatic Evaluation of Code Generation via Large Language Models [J].
Yang, Sichao ;
Yang, Ye .
2024 INTERNATIONAL SYMPOSIUM OF ELECTRONICS DESIGN AUTOMATION, ISEDA 2024, 2024, :660-665
[30]   Automatic Generation of Programming Exercises and Code Explanations Using Large Language Models [J].
Sarsa, Sami ;
Denny, Paul ;
Hellas, Arto ;
Leinonen, Juho .
PROCEEDINGS OF THE 2022 ACM CONFERENCE ON INTERNATIONAL COMPUTING EDUCATION RESEARCH, ICER 2022, VOL. 1, 2023, :27-43