PPTIF: Privacy-Preserving Transformer Inference Framework for Language Translation

被引:0
作者
Liu, Yanxin [1 ]
Su, Qianqian [1 ]
机构
[1] Qingdao Univ, Coll Comp Sci & Technol, Qingdao 266071, Peoples R China
关键词
Computational modeling; Transformers; Cryptography; Neural networks; Data models; Protocols; Task analysis; Homomorphic encryption; Outsourcing; Privacy; Privacy-preserving; replicated secret-sharing; secure multi-party computation; secure outsourcing; transformer; NEURAL-NETWORK INFERENCE; SYSTEM;
D O I
10.1109/ACCESS.2024.3384268
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The Transformer model has emerged as a prominent machine learning tool within the field of natural language processing. Nevertheless, running the Transformer model on resource-constrained devices presents a notable challenge that needs to be addressed. Although outsourcing services can significantly reduce the computational overhead associated with using the model, it also incurs privacy risks to the provider's proprietary model and the client's sensitive data. In this paper, we propose an efficient privacy-preserving Transformer inference framework (PPTIF) for language translation tasks based on three-party replicated secret-sharing techniques. PPTIF offers a secure approach for users to leverage Transformer-based applications, such as language translation, while maintaining the confidentiality of their original input and inference results, thereby preventing any disclosure to the cloud server. Meanwhile, PPTIF ensures robust protection for the model parameters, guaranteeing their integrity and confidentiality. In PPTIF, we design a series of interaction protocols to implement the secure computation of Transformer components, namely secure Encoder and secure Decoder. To improve the efficiency of PPTIF, we optimize the computation of the Scaled Dot-Product Attention (Transformer's core operation) under secret sharing, effectively reducing its computation and communication overhead. Compared with Privformer, the optimized Masked Multi-Head Attention achieves about 1.7x lower runtime and 2.3x lower communication. In total, PPTIF achieve about 1.3x lower runtime and 1.2x lower communication. The effectiveness and security of PPTIF have been rigorously evaluated through comprehensive theoretical analysis and experimental validation.
引用
收藏
页码:48881 / 48897
页数:17
相关论文
共 46 条
  • [1] A Survey on Homomorphic Encryption Schemes: Theory and Implementation
    Acar, Abbas
    Aksu, Hidayet
    Uluagac, A. Selcuk
    Conti, Mauro
    [J]. ACM COMPUTING SURVEYS, 2018, 51 (04)
  • [2] Aharoni R, 2019, Arxiv, DOI arXiv:1903.00089
  • [3] Privformer: Privacy-preserving Transformer with MPC
    Akimoto, Yoshimasa
    Fukuchi, Kazuto
    Akimoto, Youhei
    Sakuma, Jun
    [J]. 2023 IEEE 8TH EUROPEAN SYMPOSIUM ON SECURITY AND PRIVACY, EUROS&P, 2023, : 392 - 410
  • [4] [Anonymous], 2023, Multi-Protocol SPDZ
  • [5] High-Throughput Semi-Honest Secure Three-Party Computation with an Honest Majority
    Araki, Toshinori
    Furukawa, Jun
    Lindell, Yehuda
    Nof, Ariel
    Ohara, Kazuma
    [J]. CCS'16: PROCEEDINGS OF THE 2016 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, 2016, : 805 - 817
  • [6] Atallah Mikhail, 2004, P 2004 ACM WORKSH PR, P103, DOI DOI 10.1145/1029179.1029204
  • [7] Basu P, 2022, Arxiv, DOI arXiv:2106.13973
  • [8] Brown TB, 2020, ADV NEUR IN, V33
  • [9] Universally composable security: A new paradigm for cryptographic protocols
    Canetti, R
    [J]. 42ND ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, PROCEEDINGS, 2001, : 136 - 145
  • [10] Security and composition of multiparty cryptographic protocols
    Canetti, R
    [J]. JOURNAL OF CRYPTOLOGY, 2000, 13 (01) : 143 - 202