Performance Analysis of Llama 2 Among Other LLMs

被引:4
作者
Huang, Donghao [1 ,2 ]
Hu, Zhenda [3 ]
Wang, Zhaoxia [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, Singapore, Singapore
[2] Mastercard, Res & Dev, Singapore, Singapore
[3] Shanghai Univ Finance & Econ, Sch Informat Management & Engn, Shanghai, Peoples R China
来源
2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024 | 2024年
关键词
large language model; in-context learning; generative pre-trained transformer; model evaluation;
D O I
10.1109/CAI59869.2024.00108
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Llama 2, an open-source large language model developed by Meta, offers a versatile and high-performance solution for natural language processing, boasting a broad scale, competitive dialogue capabilities, and open accessibility for research and development, thus driving innovation in AI applications. Despite these advancements, there remains a limited understanding of the underlying principles and performance of Llama 2 compared with other LLMs. To address this gap, this paper presents a comprehensive evaluation of Llama 2, focusing on its application in in-context learning - an AI design pattern that harnesses pre-trained LLMs for processing confidential and sensitive data. Through a rigorous comparative analysis with other open-source LLMs and OpenAI models, this study sheds light on Llama 2's performance, quality, and potential use cases. Our findings indicate that Llama 2 holds significant promise for applications involving in-context learning, with notable strengths in both answer quality and inference speed. This research offers valuable insights for the fields of LLMs and serves as an effective reference for companies and individuals utilizing such large models. The source codes and datasets of this paper are accessible at https://github.com/inflaton/Llama-2-eval.
引用
收藏
页码:1081 / 1085
页数:5
相关论文
共 16 条
[1]  
Adlakha V, 2024, Arxiv, DOI arXiv:2307.16877
[2]  
Cai S., 2022, arXiv, DOI DOI 10.48550/ARXIV.2205.03767
[3]  
Chen Y, 2023, Arxiv, DOI arXiv:2304.00723
[4]  
Dey S, 2022, 2022 ACM/IEEE 44TH INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING: NEW IDEAS AND EMERGING RESULTS (ICSE-NIER 2022), P31, DOI [10.1145/3510455.3512790, 10.1109/ICSE-NIER55298.2022.9793517]
[5]  
Huang D., 2024, 28 PAC AS C KNOWL DI
[6]   DeepSumm: Exploiting topic models and sequence to sequence networks for extractive text summarization [J].
Joshi, Akanksha ;
Fidalgo, Eduardo ;
Alegre, Enrique ;
Fernandez-Robles, Laura .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 211
[7]  
Kamalloo E, 2023, Arxiv, DOI arXiv:2305.06984
[8]  
Lin C-Y, 2004, Text Summarization Branches Out, P74
[9]  
Min SW, 2022, Arxiv, DOI [arXiv:2202.12837, DOI 10.48550/ARXIV.2202.12837]
[10]  
Sharir O, 2020, Arxiv, DOI [arXiv:2004.08900, DOI 10.48550/ARXIV.2004.08900]