CodeT5+: Open Code Large Language Models for Code Understanding and Generation

被引:0
作者
Wang, Yue [1 ]
Le, Hung [1 ]
Gotmare, Akhilesh Deepak [1 ]
Bui, Nghi D. Q. [1 ]
Li, Junnan [1 ]
Hoi, Steven C. H. [1 ]
机构
[1] Salesforce AI Res, San Francisco, CA 94105 USA
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) pretrained on vast source code have achieved prominent progress in code intelligence. However, existing code LLMs have two main limitations. First, they often adopt a specific architecture (encoder-only or decoder-only) or rely on a unified encoder-decoder network for different downstream tasks, lacking the flexibility to operate in the optimal architecture for a specific task. Secondly, they often employ a limited set of pretraining objectives which might not be relevant to some tasks and hence result in substantial performance degrade. To address these limitations, we propose "CodeT5+", a family of encoder-decoder LLMs for code in which component modules can be flexibly combined to suit a wide range of code tasks. Such flexibility is enabled by our proposed mixture of pretraining objectives, which cover span denoising, contrastive learning, text-code matching, and causal LM pretraining tasks, on both unimodal and bimodal multilingual code corpora. Furthermore, we propose to initialize CodeT5+ with frozen off-the-shelf LLMs without training from scratch to efficiently scale up our models, and explore instruction-tuning to align with natural language instructions. We extensively evaluate CodeT5+ on over 20 code-related benchmarks in different settings, including zero-shot, finetuning, and instruction-tuning. We observe state-of-the-art (SoTA) performance on various code-related tasks, and our instruction-tuned CodeT5+ 16B achieves new SoTA results of 35.0% pass@1 and 54.5% pass@10 on the HumanEval code generation task against other open code LLMs, even surpassing the OpenAI code-cushman-001 model.
引用
收藏
页码:1069 / 1088
页数:20
相关论文
共 50 条
[31]   Synthetic Data Generation Using Large Language Models: Advances in Text and Code [J].
Nadas, Mihai ;
Diosan, Laura ;
Tomescu, Andreea .
IEEE ACCESS, 2025, 13 :134615-134633
[32]   Hot or Cold? Adaptive Temperature Sampling for Code Generation with Large Language Models [J].
Zhu, Yuqi ;
Li, Jia ;
Li, Ge ;
Zhao, YunFei ;
Li, Jia ;
Jin, Zhi ;
Mei, Hong .
THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 1, 2024, :437-445
[33]   Evaluating quantized Large Language Models for code generation on low-resource language benchmarks [J].
Nyamsuren, Enkhbold .
Journal of Computer Languages, 2025, 84
[34]   VeriGen: A Large Language Model for Verilog Code Generation [J].
Thakur, Shailja ;
Ahmad, Baleegh ;
Pearce, Hammond ;
Tan, Benjamin ;
Dolan-Gavitt, Brendan ;
Karri, Ramesh ;
Garg, Siddharth .
ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (03)
[35]   Structural language models for any-code generation [J].
Alon, Uri ;
Sadaka, Roy ;
Levy, Omer ;
Yahav, Eran .
arXiv, 2019,
[36]   Type-Constrained Code Generation with Language Models [J].
Mundler, Niels ;
He, Jingxuan ;
Wang, Hao ;
Sen, Koushik ;
Song, Dawn ;
Vechev, Martin .
PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2025, 9 (PLDI)
[37]   JavaBench: A Benchmark of Object-Oriented Code Generation for Evaluating Large Language Models [J].
Cao, Jialun ;
Chen, Zhiyong ;
Wu, Jiarong ;
Cheung, Shing-Chi ;
Xu, Chang .
Proceedings - 2024 39th ACM/IEEE International Conference on Automated Software Engineering, ASE 2024, :870-882
[38]   Using Large Language Models for Aerospace Code Generation: Methods, Benchmarks, and Potential Values [J].
He, Rui ;
Zhang, Liang ;
Lyu, Mengyao ;
Lyu, Liangqing ;
Xue, Changbin .
AEROSPACE, 2025, 12 (06)
[39]   Enhancing Large Language Models-Based Code Generation by Leveraging Genetic Improvement [J].
Pinna, Giovanni ;
Ravalico, Damiano ;
Rovito, Luigi ;
Manzoni, Luca ;
De Lorenzo, Andrea .
GENETIC PROGRAMMING, EUROGP 2024, 2024, 14631 :108-124
[40]   Can ChatGPT Support Developers? An Empirical Evaluation of Large Language Models for Code Generation [J].
Jin, Kailun ;
Wang, Chung-Yu ;
Hung Viet Pham ;
Hemmati, Hadi .
2024 IEEE/ACM 21ST INTERNATIONAL CONFERENCE ON MINING SOFTWARE REPOSITORIES, MSR, 2024, :167-171