CodeT5+: Open Code Large Language Models for Code Understanding and Generation

被引:0
作者
Wang, Yue [1 ]
Le, Hung [1 ]
Gotmare, Akhilesh Deepak [1 ]
Bui, Nghi D. Q. [1 ]
Li, Junnan [1 ]
Hoi, Steven C. H. [1 ]
机构
[1] Salesforce AI Res, San Francisco, CA 94105 USA
来源
2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023 | 2023年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Large language models (LLMs) pretrained on vast source code have achieved prominent progress in code intelligence. However, existing code LLMs have two main limitations. First, they often adopt a specific architecture (encoder-only or decoder-only) or rely on a unified encoder-decoder network for different downstream tasks, lacking the flexibility to operate in the optimal architecture for a specific task. Secondly, they often employ a limited set of pretraining objectives which might not be relevant to some tasks and hence result in substantial performance degrade. To address these limitations, we propose "CodeT5+", a family of encoder-decoder LLMs for code in which component modules can be flexibly combined to suit a wide range of code tasks. Such flexibility is enabled by our proposed mixture of pretraining objectives, which cover span denoising, contrastive learning, text-code matching, and causal LM pretraining tasks, on both unimodal and bimodal multilingual code corpora. Furthermore, we propose to initialize CodeT5+ with frozen off-the-shelf LLMs without training from scratch to efficiently scale up our models, and explore instruction-tuning to align with natural language instructions. We extensively evaluate CodeT5+ on over 20 code-related benchmarks in different settings, including zero-shot, finetuning, and instruction-tuning. We observe state-of-the-art (SoTA) performance on various code-related tasks, and our instruction-tuned CodeT5+ 16B achieves new SoTA results of 35.0% pass@1 and 54.5% pass@10 on the HumanEval code generation task against other open code LLMs, even surpassing the OpenAI code-cushman-001 model.
引用
收藏
页码:1069 / 1088
页数:20
相关论文
共 50 条
[41]   A Comprehensive Framework for Evaluating API-oriented Code Generation in Large Language Models [J].
Wu, Yixi ;
He, Pengfei ;
Wang, Zehao ;
Wang, Shaowei ;
Tian, Yuan ;
Chen, Tse-Hsun .
arXiv,
[42]   Balancing Security and Correctness in Code Generation: An Empirical Study on Commercial Large Language Models [J].
Black, Gavin S. ;
Rimal, Bhaskar P. ;
Vaidyan, Varghese Mathew .
IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025, 9 (01) :419-430
[43]   Evaluating Large Language Models in Code Generation: INFINITE Methodology for Defining the Inference Index [J].
Christakis, Nicholas ;
Drikakis, Dimitris .
APPLIED SCIENCES-BASEL, 2025, 15 (07)
[44]   VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation [J].
Vijayaraghavan, Prashanth ;
Shi, Luyao ;
Ambrogio, Stefano ;
Mackin, Charles ;
Nitsure, Apoorva ;
Beymer, David ;
Degan, Ehsan .
2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
[45]   Can Large Language Models Comprehend Code Stylometry? [J].
Dipongkor, Atish Kumar .
PROCEEDINGS OF 2024 39TH ACM/IEEE INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2024, 2024, :2429-2431
[46]   Reasoning and Planning with Large Language Models in Code Development [J].
Ding, Hao ;
Fan, Ziwei ;
Guehring, Ingo ;
Gupta, Gaurav ;
Ha, Wooseok ;
Huan, Jun ;
Liu, Linbo ;
Omidvar-Tehrani, Behrooz ;
Wang, Shiqi ;
Zhou, Hao .
PROCEEDINGS OF THE 30TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2024, 2024, :6480-6490
[47]   Code Comprehension: Review and Large Language Models Exploration [J].
Cui, Jielun ;
Zhao, Yutong ;
Yu, Chong ;
Huang, Jiaqi ;
Wu, Yuanyuan ;
Zhao, Yu .
2024 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE, SEAI 2024, 2024, :183-187
[48]   Can large language models generate geospatial code? [J].
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China ;
不详 .
arXiv, 1600,
[49]   Code Soliloquies for Accurate Calculations in Large Language Models [J].
Sonkar, Shashank ;
Chen, Xinghe ;
Le, MyCo ;
Liu, Naiming ;
Mallick, Debshila Basu ;
Baraniuk, Richard G. .
FOURTEENTH INTERNATIONAL CONFERENCE ON LEARNING ANALYTICS & KNOWLEDGE, LAK 2024, 2024, :828-835
[50]   Can Large Language Models Write Parallel Code? [J].
Nichols, Daniel ;
Davis, Joshua H. ;
Xie, Zhaojun ;
Rajaram, Arjun ;
Bhatele, Abhinav .
PROCEEDINGS OF THE 33RD INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2024, 2024,