Privet: A Privacy-Preserving Vertical Federated Learning Service for Gradient Boosted Decision Tables

被引:6
作者
Zheng Y. [1 ]
Xu S. [1 ]
Wang S. [1 ]
Gao Y. [2 ]
Hua Z. [1 ]
机构
[1] Harbin Institute of Technology, School of Computer Science and Technology, Guangdong, Shenzhen
[2] Csiro, Data61, Sydney, 2720, NSW
来源
IEEE Transactions on Services Computing | 2023年 / 16卷 / 05期
关键词
decision table; gradient boosting; multi-party collaboration; privacy preservation; Vertical federated learning service;
D O I
10.1109/TSC.2023.3279839
中图分类号
学科分类号
摘要
Vertical federated learning (VFL) has recently emerged as an appealing distributed paradigm empowering multi-party collaboration for training high-quality models over vertically partitioned datasets. Gradient boosting has been popularly adopted in VFL, which builds an ensemble of weak learners (typically decision trees) to achieve promising prediction performance. Recently there have been growing interests in using decision table as an intriguing alternative weak learner in gradient boosting, due to its simpler structure, good interpretability, and promising performance. In the literature, there have been works on privacy-preserving VFL for gradient boosted decision trees, but no prior work has been devoted to the emerging case of decision tables. Training and inference on decision tables are different from that in the case of generic decision trees, not to mention gradient boosting with decision tables in VFL. In light of this, we design, implement, and evaluate Privet, the first system framework enabling privacy-preserving VFL service for gradient boosted decision tables. Privet delicately builds on lightweight cryptography and allows an arbitrary number of participants holding vertically partitioned datasets to securely train gradient boosted decision tables. Extensive experiments over several real-world datasets and synthetic datasets demonstrate that Privet achieves promising performance, with utility comparable to plaintext centralized learning. © 2008-2012 IEEE.
引用
收藏
页码:3604 / 3620
页数:16
相关论文
共 56 条
  • [1] Bonawitz K., Et al., Practical secure aggregation for privacy-preserving machine learning, Proc. ACM Conf. Comput. Commun. Secur, pp. 1175-1191, (2017)
  • [2] Qiao C., Brown K.N., Zhang F., Tian Z., Federated adaptive asynchronous clustering algorithm for wireless mesh networks, IEEE Trans. Knowl. Data Eng, 35, 3, pp. 2610-2627, (2023)
  • [3] Zhou P., Wang K., Guo L., Gong S., Zheng B., A privacy-preserving distributed contextual federated online learning framework with Big Data support in social recommender systems, IEEE Trans. Knowl. Data Eng, 33, 3, pp. 824-838, (2021)
  • [4] Zhao J., Et al., CORK: A privacy-preserving and lossless federated learning scheme for deep neural network, Inf. Sci, 603, pp. 190-209, (2022)
  • [5] Li Q., Wen Z., He B., Practical federated gradient boosting decision trees, Proc. Conf. Assoc. Advance. Artif. Intell, pp. 4642-4649, (2020)
  • [6] Maddock S., Cormode G., Wang T., Maple C., Jha S., Federated boosted decision trees with differential privacy, Proc. ACM Conf. Comput. Commun. Secur, pp. 2249-2263, (2022)
  • [7] Xie L., Liu J., Lu S., Chang T.-H., Shi Q., An efficient learning framework for federated XGBoost using secret sharing and distributed optimization, ACM Trans. Intell. Syst. Technol, 13, 5, pp. 1-28, (2022)
  • [8] Fang W., Et al., Large-scale secure XGB for vertical federated learning, Proc. ACM Int. Conf. Inf. Knowl. Manage, pp. 443-452, (2021)
  • [9] Fu F., Et al., VF2 boost: Very fast vertical federated gradient boosting for cross-enterprise learning, Proc. ACM Int. Conf. Manage. Data, pp. 563-576, (2021)
  • [10] Cheng K., Et al., SecureBoost: A lossless federated learning framework, IEEE Intell. Syst, 36, 6, pp. 87-98, (2021)