VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation

被引:0
作者
Vijayaraghavan, Prashanth [1 ]
Shi, Luyao [1 ]
Ambrogio, Stefano [1 ]
Mackin, Charles [1 ]
Nitsure, Apoorva [1 ]
Beymer, David [1 ]
Degan, Ehsan [1 ]
机构
[1] IBM Res, San Jose, CA 95120 USA
来源
2024 IEEE LLM AIDED DESIGN WORKSHOP, LAD 2024 | 2024年
关键词
LLMs; large language models; VHDL Code generation; VHDL Evaluation; hardware design automation; Hardware Description Languages; HDL; PEFT; ICL;
D O I
10.1109/LAD62341.2024.10691836
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the unprecedented advancements in Large Language Models (LLMs), their application domains have expanded to include code generation tasks across various programming languages. While significant progress has been made in enhancing LLMs for popular programming languages, there exists a notable gap in comprehensive evaluation frameworks tailored for Hardware Description Languages (HDLs), particularly VHDL. This paper addresses this gap by introducing a comprehensive evaluation framework designed specifically for assessing LLM performance in VHDL code generation task. We construct a dataset for evaluating LLMs on VHDL code generation task. This dataset is constructed by translating a collection of Verilog evaluation problems to VHDL and aggregating publicly available VHDL problems, resulting in a total of 202 problems. To assess the functional correctness of the generated VHDL code, we utilize a curated set of self-verifying testbenches specifically designed for those aggregated VHDL problem set. We conduct an initial evaluation of different LLMs and their variants, including zero-shot code generation, in-context learning (ICL), and Parameter-efficient fine-tuning (P EFT) methods. Our findings underscore the considerable challenges faced by existing LLMs in VHDL code generation, revealing significant scope for improvement. This study emphasizes the necessity of supervised fine-tuning code generation models specifically for VHDL, offering potential benefits to V HDL designers seeking efficient code generation solutions.
引用
收藏
页数:6
相关论文
共 22 条
[1]  
Blocklove Jason, 2023, 2023 ACM/IEEE 5th Workshop on Machine Learning for CAD (MLCAD), P1, DOI 10.1109/MLCAD58807.2023.10299874
[2]  
Chang KY, 2023, Arxiv, DOI arXiv:2305.14019
[3]  
Chen Mark., 2021, CORR, P2021, DOI [10.48550/arXiv.2107.03374, DOI 10.48550/ARXIV.2107.03374]
[4]  
DeLorenzo M, 2024, Arxiv, DOI arXiv:2402.03289
[5]  
Dettmers T, 2023, PR MACH LEARN RES, V202
[6]   Machine Learning for Electronic Design Automation: A Survey [J].
Huang, Guyue ;
Hu, Jingbo ;
He, Yifan ;
Liu, Jialong ;
Ma, Mingyuan ;
Shen, Zhaoyang ;
Wu, Juejian ;
Xu, Yuanfan ;
Zhang, Hengrui ;
Zhong, Kai ;
Ning, Xuefei ;
Ma, Yuzhe ;
Yang, Haoyu ;
Yu, Bei ;
Yang, Huazhong ;
Wang, Yu .
ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2021, 26 (05)
[7]  
Kande R, 2024, Arxiv, DOI arXiv:2306.14027
[8]  
Liu MJ, 2024, Arxiv, DOI arXiv:2311.00176
[9]  
Liu S, 2024, Arxiv, DOI arXiv:2312.08617
[10]   RTLLM: An Open-Source Benchmark for Design RTL Generation with Large Language Model [J].
Lu, Yao ;
Liu, Shang ;
Zhang, Qijun ;
Xie, Zhiyao .
29TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2024, 2024, :722-727