zkLLM: Zero Knowledge Proofs for Large Language Models

被引:1
作者
Sun, Haochen [1 ]
Li, Jason [1 ]
Zhang, Hongyang [1 ]
机构
[1] Univ Waterloo, Waterloo, ON, Canada
来源
PROCEEDINGS OF THE 2024 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2024 | 2024年
基金
加拿大自然科学与工程研究理事会;
关键词
large language models; zero-knowledge proofs;
D O I
10.1145/3658644.3670334
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The recent surge in artificial intelligence (AI), characterized by the prominence of large language models (LLMs), has ushered in fundamental transformations across the globe. However, alongside these advancements, concerns surrounding the legitimacy of LLMs have grown, posing legal challenges to their extensive applications. Compounding these concerns, the parameters of LLMs are often treated as intellectual property, restricting direct investigations. In this study, we address a fundamental challenge within the realm of AI legislation: the need to establish the authenticity of outputs generated by LLMs. To tackle this issue, we present zkLLM, which stands as the inaugural specialized zero-knowledge proof tailored for LLMs to the best of our knowledge. Addressing the persistent challenge of non-arithmetic operations in deep learning, we introduce tlookup, a parallelized lookup argument designed for non-arithmetic tensor operations in deep learning, offering a solution with no asymptotic overhead. Furthermore, leveraging the foundation of tlookup, we introduce zkAttn, a specialized zeroknowledge proof crafted for the attention mechanism, carefully balancing considerations of running time, memory usage, and accuracy. Empowered by our fully parallelized CUDA implementation, zkLLM emerges as a significant stride towards achieving efficient zero-knowledge verifiable computations over LLMs. Remarkably, for LLMs boasting 13 billion parameters, our approach enables the generation of a correctness proof for the entire inference process in under 15 minutes. The resulting proof, compactly sized at less than 200 kB, is designed to uphold the privacy of the model parameters, ensuring no inadvertent information leakage.
引用
收藏
页码:4405 / 4419
页数:15
相关论文
共 59 条
[1]  
Betti A, 2018, Arxiv, DOI arXiv:1807.06450
[2]   Ligero plus plus : A New Optimized Sublinear IOP [J].
Bhadauria, Rishabh ;
Fang, Zhiyong ;
Hazay, Carmit ;
Venkitasubramaniam, Muthuramakrishnan ;
Xie, Tiancheng ;
Zhang, Yupeng .
CCS '20: PROCEEDINGS OF THE 2020 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2020, :2025-2038
[3]   Succinct Non-Interactive Arguments via Linear Interactive Proofs [J].
Bitansky, Nir ;
Chiesa, Alessandro ;
Ishai, Yuval ;
Ostrovsky, Rafail ;
Paneth, Omer .
JOURNAL OF CRYPTOLOGY, 2022, 35 (03)
[4]  
BONEH D., 2001, Lecture Notes Comput. Sci., V2248, P514, DOI [10.1007/3-540-45682-130, DOI 10.1007/3-540-45682-130]
[5]  
Boneh Dan., IACR Cryptol. ePrint Arch, V2020, P2020
[6]  
Bowe S., 2019, IACR Crypt. ePrint Archive, P1021
[7]  
Brown TB, 2020, ADV NEUR IN, V33
[8]  
Chen Binyi, 2023, Lecture Notes in Computer Science, V14005
[9]  
Chiesa Alessandro, 2017, Electron. Colloquium Comput. Complex. TR17-057
[10]  
Chowdhery A, 2023, J MACH LEARN RES, V24