Planter: Rapid Prototyping of In-Network Machine Learning Inference

被引:0
作者
Zheng, Changgang [1 ]
Zang, Mingyuan [2 ]
Hong, Xinpeng [1 ]
Perreault, Liam [1 ]
Bensoussane, Riyad [1 ]
Vargaftik, Shay [3 ]
Ben-Itzhak, Yaniv [3 ]
Zilberman, Noa [1 ]
机构
[1] Univ Oxford, Oxford, England
[2] Tech Univ Denmark, Kongens Lyngby, Denmark
[3] VMware Res, Broadcom, Palo Alto, CA USA
关键词
In-Network Computing; Machine Learning; Dimension Reduction; Modular Framework; Machine Learning Compilers; Programmable Switches; P4;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In-network machine learning inference provides high throughput and low latency. It is ideally located within the network, power efficient, and improves applications' performance. Despite its advantages, the bar to in-network machine learning research is high, requiring significant expertise in programmable data planes, in addition to knowledge of machine learning and the application area. Existing solutions are mostly one-time efforts, hard to reproduce, change, or port across platforms. In this paper, we present Planter: a modular and efficient open-source framework for rapid prototyping of in-network machine learning models across a range of platforms and pipeline architectures. By identifying general mapping methodologies for machine learning algorithms, Planter introduces new machine learning mappings and improves existing ones. It provides users with several example use cases and supports different datasets, and was already extended by users to new fields and applications. Our evaluation shows that Planter improves machine learning performance compared with previous model-tailored works, while significantly reducing resource consumption and co-existing with network functionality. Planter-supported algorithms run at line rate on unmodified commodity hardware, providing billions of inference decisions per second.
引用
收藏
页码:2 / 20
页数:19
相关论文
共 106 条
[91]   Do Switches Dream of Machine Learning? Toward In-Network Classification [J].
Xiong, Zhaoqi ;
Zilberman, Noa .
PROCEEDINGS OF THE EIGHTEENTH ACM WORKSHOP ON HOT TOPICS IN NETWORKS (HOTNETS '19), 2019, :25-33
[92]  
Zang Mingyuan, 2022, SIGCOMM '22: Proceedings of the SIGCOMM '22 Poster and Demo Sessions, P46, DOI 10.1145/3546037.3546060
[93]  
Zang Mingyuan, 2023, Security for IoT Networks and Devices in 6G (Sec4IoT).
[94]  
Zang Mingyuan, 2023, IEEE Internet of Things Journal., V2023
[95]   pHeavy: Predicting Heavy Flows in the Programmable Data Plane [J].
Zhang, Xiaoquan ;
Cui, Lin ;
Tso, Fung Po ;
Jia, Weijia .
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2021, 18 (04) :4353-4364
[96]   A Survey on Emerging Computing Paradigms for Big Data [J].
Zhang Yaoxue ;
Ren Ju ;
Liu Jiagang ;
Xu Chugui ;
Guo Hui ;
Liu Yaping .
CHINESE JOURNAL OF ELECTRONICS, 2017, 26 (01) :1-12
[97]   DeepLOB: Deep Convolutional Neural Networks for Limit Order Books [J].
Zhang, Zihao ;
Zohren, Stefan ;
Roberts, Stephen .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (11) :3001-3012
[98]  
Zheng C., 2022, arXiv
[99]   IIsy: Hybrid In-Network Classification Using Programmable Switches [J].
Zheng, Changgang ;
Xiong, Zhaoqi ;
Bui, Thanh T. ;
Kaupmees, Siim ;
Bensoussane, Riyad ;
Bernabeu, Antoine ;
Vargaftik, Shay ;
Ben-Itzhak, Yaniv ;
Zilberman, Noa .
IEEE-ACM TRANSACTIONS ON NETWORKING, 2024, 32 (03) :2555-2570
[100]   In-Network Machine Learning Using Programmable Network Devices: A Survey [J].
Zheng, Changgang ;
Hong, Xinpeng ;
Ding, Damu ;
Vargaftik, Shay ;
Ben-Itzhak, Yaniv ;
Zilberman, Noa .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2024, 26 (02) :1171-1200