Large Language Models and Simple, Stupid Bugs

被引:11
|
作者
Jesse, Kevin [1 ]
Ahmed, Toufique [1 ]
Devanbu, Premkumar T. [1 ]
Morgan, Emily [1 ]
机构
[1] Univ Calif Davis, Davis, CA 95616 USA
来源
2023 IEEE/ACM 20TH INTERNATIONAL CONFERENCE ON MINING SOFTWARE REPOSITORIES, MSR | 2023年
基金
美国国家科学基金会;
关键词
language models; prompting; deep learning; software engineering;
D O I
10.1109/MSR59073.2023.00082
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
With the advent of powerful neural language models, AI-based systems to assist developers in coding tasks are becoming widely available; Copilot is one such system. Copilot uses Codex, a large language model (LLM), to complete code conditioned on a preceding "prompt". Codex, however, is trained on public GitHub repositories, viz., on code that may include bugs and vulnerabilities. Previous studies [1], [2] show Codex reproduces vulnerabilities seen in training. In this study, we examine how prone Codex is to generate an interesting bug category, single statement bugs, commonly referred to as simple, stupid bugs or SStuBs in the MSR community. We find that Codex and similar LLMs do help avoid some SStuBs, but do produce known, verbatim SStuBs as much as 2x as likely than known, verbatim correct code. We explore the consequences of the Codex generated SStuBs and propose avoidance strategies that suggest the possibility of reducing the production of known, verbatim SStubs, and increase the possibility of producing known, verbatim fixes.
引用
收藏
页码:563 / 575
页数:13
相关论文
共 50 条
  • [41] Enhancing Urban Walkability Assessment with Multimodal Large Language Models
    Blecic, Ivan
    Saiu, Valeria
    Trunfio, Giuseppe A.
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS-ICCSA 2024 WORKSHOPS, PT V, 2024, 14819 : 394 - 411
  • [42] AskIt: Unified Programming Interface for Programming with Large Language Models
    Okuda, Katsumi
    Amarasinghe, Saman
    2024 IEEE/ACM INTERNATIONAL SYMPOSIUM ON CODE GENERATION AND OPTIMIZATION, CGO, 2024, : 41 - 54
  • [43] Large language models for code completion: A systematic literature review
    Husein, Rasha Ahmad
    Aburajouh, Hala
    Catal, Cagatay
    COMPUTER STANDARDS & INTERFACES, 2025, 92
  • [44] Persistent Anti-Muslim Bias in Large Language Models
    Abid, Abubakar
    Farooqi, Maheen
    Zou, James
    AIES '21: PROCEEDINGS OF THE 2021 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2021, : 298 - 306
  • [45] Measuring and Improving the Energy Efficiency of Large Language Models Inference
    Argerich, Mauricio Fadel
    Patino-Martinez, Marta
    IEEE ACCESS, 2024, 12 : 80194 - 80207
  • [46] Investigating legal question generation using large language models
    Deroy, Aniket
    Ghosh, Kripabandhu
    Ghosh, Saptarshi
    ARTIFICIAL INTELLIGENCE AND LAW, 2025,
  • [47] QueryMintAI: Multipurpose Multimodal Large Language Models for Personal Data
    Ghosh, Ananya
    Deepa, K.
    IEEE ACCESS, 2024, 12 : 144631 - 144651
  • [48] Comparative Analysis of Large Language Models in Source Code Analysis
    Erdogan, Huseyin
    Turan, Nezihe Turhan
    Onan, Aytug
    INTELLIGENT AND FUZZY SYSTEMS, INFUS 2024 CONFERENCE, VOL 1, 2024, 1088 : 185 - 192
  • [49] CIRCUITSYNTH: Leveraging Large Language Models for Circuit Topology Synthesis
    Vijayaraghavan, Prashanth
    Shi, Luyao
    Degan, Ehsan
    Zhang, Xin
    2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024, 2024,
  • [50] How Useful Are Educational Questions Generated by Large Language Models?
    Elkins, Sabina
    Kochmar, Ekaterina
    Serban, Iulian
    Cheung, Jackie C. K.
    ARTIFICIAL INTELLIGENCE IN EDUCATION. POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS, DOCTORAL CONSORTIUM AND BLUE SKY, AIED 2023, 2023, 1831 : 536 - 542