Large language models for code completion: A systematic literature review

被引:0
|
作者
Husein, Rasha Ahmad [1 ]
Aburajouh, Hala [1 ]
Catal, Cagatay [1 ]
机构
[1] Qatar Univ, Dept Comp Sci & Engn, Doha, Qatar
关键词
Code completion; Large language models; Deep learning; Transformers;
D O I
10.1016/j.csi.2024.103917
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Code completion serves as a fundamental aspect of modern software development, improving developers' coding processes. Integrating code completion tools into an Integrated Development Environment (IDE) or code editor enhances the coding process and boosts productivity by reducing errors and speeding up code writing while reducing cognitive load. This is achieved by predicting subsequent tokens, such as keywords, variable names, types, function names, operators, and more. Different techniques can achieve code completion, and recent research has focused on Deep Learning methods, particularly Large Language Models (LLMs) utilizing Transformer algorithms. While several research papers have focused on the use of LLMs for code completion, these studies are fragmented, and there is no systematic overview of the use of LLMs for code completion. Therefore, we aimed to perform a Systematic Literature Review (SLR) study to investigate how LLMs have been applied for code completion so far. We have formulated several research questions to address how LLMs have been integrated for code completion-related tasks and to assess the efficacy of these LLMs in the context of code completion. To achieve this, we retrieved 244 papers from scientific databases using auto-search and specific keywords, finally selecting 23 primary studies based on an SLR methodology for in-depth analysis. This SLR study categorizes the granularity levels of code completion achieved by utilizing LLMs in IDEs, explores the existing issues in current code completion systems, how LLMs address these challenges, and the pre-training and fine-tuning methods employed. Additionally, this study identifies open research problems and outlines future research directions. Our analysis reveals that LLMs significantly enhance code completion performance across several programming languages and contexts, and their capability to predict relevant code snippets based on context and partial input boosts developer productivity substantially.
引用
收藏
页数:15
相关论文
共 50 条
  • [1] Large Language Models for Software Engineering: A Systematic Literature Review
    Hou, Xinyi
    Zhao, Yanjie
    Liu, Yue
    Yang, Zhou
    Wang, Kailong
    Li, Li
    Luo, Xiapu
    Lo, David
    Grundy, John
    Wang, Haoyu
    ACM Transactions on Software Engineering and Methodology, 2024, 33 (08)
  • [2] Application of Large Language Models in Cybersecurity: A Systematic Literature Review
    Hasanov, Ismayil
    Virtanen, Seppo
    Hakkala, Antti
    Isoaho, Jouni
    IEEE ACCESS, 2024, 12 : 176751 - 176778
  • [3] Large Language Models in Computer Science Education: A Systematic Literature Review
    Raihan, Nishat
    Siddiq, Mohammed Latif
    Santos, Joanna C. S.
    Zampieri, Marcos
    PROCEEDINGS OF THE 56TH ACM TECHNICAL SYMPOSIUM ON COMPUTER SCIENCE EDUCATION, SIGCSE TS 2025, VOL 2, 2025, : 938 - 944
  • [4] Large Language Models in Computer Science Education: A Systematic Literature Review
    Raihan, Nishat
    Siddiq, Mohammed Latif
    Santos, Joanna C. S.
    Zampieri, Marcos
    PROCEEDINGS OF THE 56TH ACM TECHNICAL SYMPOSIUM ON COMPUTER SCIENCE EDUCATION, SIGCSE TS 2025, VOL 1, 2025, : 938 - 944
  • [5] Code Completion with Statistical Language Models
    Raychev, Veselin
    Vechev, Martin
    Yahav, Eran
    ACM SIGPLAN NOTICES, 2014, 49 (06) : 419 - 428
  • [6] A COMPARATIVE ANALYSIS OF LARGE LANGUAGE MODELS (LLM) UTILISED IN SYSTEMATIC LITERATURE REVIEW
    Rathi, H.
    Malik, A.
    Behera, D. C.
    Kamboj, G.
    VALUE IN HEALTH, 2023, 26 (12) : S6 - S6
  • [7] Code Comprehension: Review and Large Language Models Exploration
    Cui, Jielun
    Zhao, Yutong
    Yu, Chong
    Huang, Jiaqi
    Wu, Yuanyuan
    Zhao, Yu
    2024 IEEE 4TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING AND ARTIFICIAL INTELLIGENCE, SEAI 2024, 2024, : 183 - 187
  • [8] From Large Language Models to Large Multimodal Models: A Literature Review
    Huang, Dawei
    Yan, Chuan
    Li, Qing
    Peng, Xiaojiang
    APPLIED SCIENCES-BASEL, 2024, 14 (12):
  • [9] Large Language Models and Empathy: Systematic Review
    Sorin, Vera
    Brin, Dana
    Barash, Yiftach
    Konen, Eli
    Charney, Alexander
    Nadkarni, Girish
    Klang, Eyal
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26
  • [10] Large Language Models in Gastroenterology: Systematic Review
    Gong, Eun Jeong
    Bang, Chang Seok
    Lee, Jae Jun
    Park, Jonghyung
    Kim, Eunsil
    Kim, Subeen
    Kimm, Minjae
    Choi, Seoung-Ho
    JOURNAL OF MEDICAL INTERNET RESEARCH, 2024, 26