Memory Network Architecture for Packet Processing in Functions Virtualization

被引:0
作者
Korikawa, Tomohiro [1 ,2 ]
Oki, Eiji [1 ]
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto 6068501, Japan
[2] NTT Corp, Network Innovat Ctr, Tokyo 1808585, Japan
来源
IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT | 2022年 / 19卷 / 03期
关键词
Memory management; Random access memory; Computer architecture; Task analysis; Virtualization; Hardware; Performance evaluation; Network functions virtualization; packet processing; memory network; memory-centric computing; disaggregated computing; DRAM; P4;
D O I
10.1109/TNSM.2022.3159091
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Packet processing tasks in network functions require high-performance memory systems to understand the packet information, update the packet content, and search the databases. While network virtualization is expected to bring flexible and adaptive network with reduced cost by using commercial off-the-shelf (COTS) hardware and programmable data plane technology, network function performance suffers from the poor memory systems in COTS computers and lack of scalability in programmable hardware devices. This paper proposes a memory network architecture for packet processing based on memory-centric, disaggregated computing. Unlike processor-centric architecture in today's COTS computers, the memory network consists of multiple memory devices, where processing for the incoming packets is completed. The proposed architecture reduces packet processing latency by eliminating communication between the processor devices and the memory devices. Also, the proposed architecture provides scalability of hardware resources by dynamic memory device allocation depending on the complexity of the network function, memory-intensiveness of packet processing, and traffic load. The numerical results show that the proposed architecture reduces accumulated latency for memory accesses and increases throughput compared to the conventional, processor-centric architectures, where every memory access requires communication between the processor devices and the memory devices. The proposed architecture also reduces latency and increases throughput by allocating additional memory devices to memory-intensive tasks.
引用
收藏
页码:3304 / 3322
页数:19
相关论文
共 93 条
  • [1] PIM-Enabled Instructions: A Low-Overhead, Locality-Aware Processing-in-Memory Architecture
    Ahn, Junwhan
    Yoo, Sungjoo
    Mutlu, Onur
    Choi, Kiyoung
    [J]. 2015 ACM/IEEE 42ND ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2015, : 336 - 348
  • [2] Alves MAZ, 2016, DES AUT TEST EUROPE, P1249
  • [3] [Anonymous], PCI SIG SINGLE ROOT
  • [4] [Anonymous], 2018, CISC VIS NETW IND GL
  • [5] [Anonymous], 3GPP RELEASES
  • [6] [Anonymous], INTEL DATA PLANE DEV
  • [7] Poptrie: A Compressed Trie with Population Count for Fast and Scalable Software IP Routing Table Lookup
    Asai, Hirochika
    Ohara, Yasuhiro
    [J]. ACM SIGCOMM COMPUTER COMMUNICATION REVIEW, 2015, 45 (04) : 57 - 70
  • [8] Bando M, 2010, IEEE INFOCOM SER
  • [9] High-Speed Software Data Plane via Vectorized Packet Processing
    Barach, David
    Linguaglossa, Leonardo
    Marion, Damjan
    Pfister, Pierre
    Pontarelli, Salvatore
    Rossi, Dario
    [J]. IEEE COMMUNICATIONS MAGAZINE, 2018, 56 (12) : 97 - 103
  • [10] Bedeschi Ferdinando, 2008, 2008 IEEE International Solid-State Circuits Conference - Digest of Technical Papers, P428, DOI 10.1109/ISSCC.2008.4523240