Large Language Models and Simple, Stupid Bugs

被引:11
|
作者
Jesse, Kevin [1 ]
Ahmed, Toufique [1 ]
Devanbu, Premkumar T. [1 ]
Morgan, Emily [1 ]
机构
[1] Univ Calif Davis, Davis, CA 95616 USA
来源
2023 IEEE/ACM 20TH INTERNATIONAL CONFERENCE ON MINING SOFTWARE REPOSITORIES, MSR | 2023年
基金
美国国家科学基金会;
关键词
language models; prompting; deep learning; software engineering;
D O I
10.1109/MSR59073.2023.00082
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
With the advent of powerful neural language models, AI-based systems to assist developers in coding tasks are becoming widely available; Copilot is one such system. Copilot uses Codex, a large language model (LLM), to complete code conditioned on a preceding "prompt". Codex, however, is trained on public GitHub repositories, viz., on code that may include bugs and vulnerabilities. Previous studies [1], [2] show Codex reproduces vulnerabilities seen in training. In this study, we examine how prone Codex is to generate an interesting bug category, single statement bugs, commonly referred to as simple, stupid bugs or SStuBs in the MSR community. We find that Codex and similar LLMs do help avoid some SStuBs, but do produce known, verbatim SStuBs as much as 2x as likely than known, verbatim correct code. We explore the consequences of the Codex generated SStuBs and propose avoidance strategies that suggest the possibility of reducing the production of known, verbatim SStubs, and increase the possibility of producing known, verbatim fixes.
引用
收藏
页码:563 / 575
页数:13
相关论文
共 50 条
  • [1] Large Language Models are Not Models of Natural Language: They are Corpus Models
    Veres, Csaba
    IEEE ACCESS, 2022, 10 : 61970 - 61979
  • [2] Unveiling the potential of large language models in generating semantic and cross-language clones
    Roy, Palash R.
    Alam, Ajmain I.
    Al-omari, Farouq
    Roy, Banani
    Roy, Chanchal K.
    Schneider, Kevin A.
    2023 IEEE 17TH INTERNATIONAL WORKSHOP ON SOFTWARE CLONES, IWSC 2023, 2023, : 22 - 28
  • [3] Symbols and grounding in large language models
    Pavlick, Ellie
    PHILOSOPHICAL TRANSACTIONS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2023, 381 (2251):
  • [4] Can Large Language Models Replace Therapists? Evaluating Performance at Simple Cognitive Behavioral Therapy Tasks
    Hodson, Nathan
    Williamson, Simon
    JMIR AI, 2024, 3
  • [5] Prompting Large Language Models With the Socratic Method
    Chang, Edward Y.
    2023 IEEE 13TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE, CCWC, 2023, : 351 - 360
  • [6] The use of large language models for program repair
    Zubair, Fida
    Al-Hitmi, Maryam
    Catal, Cagatay
    COMPUTER STANDARDS & INTERFACES, 2025, 93
  • [7] Assessing the Strengths and Weaknesses of Large Language Models
    Shalom Lappin
    Journal of Logic, Language and Information, 2024, 33 : 9 - 20
  • [8] Assessing political bias in large language models
    Rettenberger, Luca
    Reischl, Markus
    Schutera, Mark
    JOURNAL OF COMPUTATIONAL SOCIAL SCIENCE, 2025, 8 (02):
  • [9] Level Generation Through Large Language Models
    Todd, Graham
    Earle, Sam
    Nasir, Muhammad Umair
    Green, Michael Cerny
    Togelius, Julian
    PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON THE FOUNDATIONS OF DIGITAL GAMES, FDG 2023, 2023,
  • [10] Applying Large Language Models to Issue Classification
    Aracena, Gabriel
    Luster, Kyle
    Santos, Fabio
    Steinmacher, Igor
    Gerosa, Marco Aurelio
    PROCEEDINGS 2024 ACM/IEEE INTERNATIONAL WORKSHOP ON NL-BASED SOFTWARE ENGINEERING, NLBSE 2024, 2024, : 57 - 60