The High Performance Interconnect Architecture for Supercomputers

被引:0
作者
Simonov A.S. [1 ,2 ]
Semenov A.S. [1 ]
Shcherbak A.N. [1 ]
Zhabin I.A. [1 ]
机构
[1] Federal State Unitary Enterprise “Russian Federal Nuclear Center-Zababakhin All – Russia Research Institute of Technical Physics”, Snezhinsk
[2] Federal State Budgetary Educational Institution of Higher Education Moscow Aviation Institute (National Research University), Moscow
关键词
Angara; high performance computing; interconnect; supercomputer;
D O I
10.14529/JSFI230208
中图分类号
学科分类号
摘要
In this paper, we introduce the design of an advanced high-performance interconnect architecture for supercomputers. In the first part of the paper, we consider the first generation high-performance Angara interconnect (Angara G1). The Angara interconnect is based on the router ASIC, which supports a 4D torus topology, a deterministic and an adaptive routing, and has the hardware support of the RDMA technology. The interface with a processor unit is PCI Express. The Angara G1 interconnect has an extremely low communication latency of 850 ns using the MPI library, as well as a link bandwidth of 75 Gbps. In the paper, we present the scalability performance results of the considered application problems on the supercomputers equipped with the Angara G1 interconnect. In the second part of the paper, using research results and experience we present the architecture of the advanced interconnect for supercomputers (G2). The G2 architecture supports 6D torus topology, the advanced deterministic and zone adaptive routing algorithms, and a low-level interconnect operations including acknowledgments and notifications. G2 includes support for exceptions, performance counters, and SR-IOV virtualization. A G2 hardware is planned in the form factor of a 32-port switch with the QSFP-DD connectors and a two-port low profile PCI Express adapter. The switches can be combined to 4D torus topology. We show the performance evaluation of an experimental FPGA prototype, which confirm the possibility of implementing the proposed advanced high performance interconnect architecture. © The Authors 2023. This paper is published with open access at SuperFri.org
引用
收藏
页码:127 / 136
页数:9
相关论文
共 24 条
[1]  
Abts D., The Cray XT4 and Seastar 3D torus interconnect, (2011)
[2]  
Adiga N.R., Blumrich M.A., Chen D., Et al., Blue Gene/L torus interconnection network, IBM Journal of Research and Development, 49, 2, pp. 265-276, (2005)
[3]  
Agarkov A., Ismagilov T., Makagon D., Et al., Performance evaluation of the Angara interconnect, Proc. Int. Conf. on Russian Supercomputing Days, pp. 626-639, (2016)
[4]  
Akimov V., Silaev D., Aksenov A., Et al., FlowVision scalability on supercomputers with Angara interconnect, Lobachevskii Journal of Mathematics, 39, pp. 1159-1169, (2018)
[5]  
Alam S.R., Kuehn J.A., Barrett R.F., Et al., Cray XT4: an early evaluation for petascale scientific simulation, Proceedings of the 2007 ACM/IEEE Conference on Supercomputing, pp. 1-12, (2007)
[6]  
Almasi G., Asaad S., Bellofatto R.E., Et al., Overview of the IBM Blue Gene/P project, IBM Journal of Research and Development, 52, 1-2, pp. 199-220, (2008)
[7]  
Alverson R., Roweth D., Kaplan L., The Gemini system interconnect, 2010 18th IEEE Symposium on High Performance Interconnects, pp. 83-87, (2010)
[8]  
Chen D., Eisley N., Heidelberger P., Et al., The IBM Blue Gene/Q interconnection fabric, IEEE Micro, 32, 1, pp. 32-43, (2011)
[9]  
Dally W.J., Seitz C.L., The torus routing chip, Distributed computing, 1, 4, pp. 187-196, (1986)
[10]  
Duato J., Yalamanchili S., Ni L., Interconnection networks, (2003)