Adaptive and Dynamic Design for MPI Tag Matching

被引:33
作者
Bayatpour, M. [1 ]
Subramoni, H. [1 ]
Chakraborty, S. [1 ]
Panda, D. K. [1 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH 43210 USA
来源
2016 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER) | 2016年
关键词
D O I
10.1109/CLUSTER.2016.69
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The Message Passing Interface (MPI) standard specifies the use of (source, tag, communicator) tuple to identify whether an incoming message is what the receiver process is expecting. The cost associated with this process, commonly known as "tag matching", is tightly coupled with the communication pattern of the application and the load it generates at each individual process. Although researchers have proposed several schemes to mitigate the cost of tag matching, they have all been static and do not adapt dynamically to the communication load at individual processes and can lead to degradation in tag matching times. Such static designs can also lead to unnecessary memory overheads for queue management at processes whose communication loads do not justify enhanced tag matching schemes. In this paper, we take up this challenge and propose a tag matching design which dynamically adapts to the communication load at each individual process at runtime. Our experimental evaluation shows that the proposed adaptive and dynamic tag matching scheme is able to deliver the best performance when compared with multiple state-of-the-art tag matching schemes while limiting the memory consumed to the absolute minimum necessary to deliver the desired performance benefits. For instance, with HPCG, the adaptive scheme delivers 20%, 32% and up to 2X improvements in tag matching performance when compared to the default, bin-based and rank-based schemes, respectively. To the best of our knowledge this is the first tag matching design that is capable of dynamically adapting to the communication requirements of end applications.
引用
收藏
页码:1 / 10
页数:10
相关论文
共 15 条
[11]  
Heroux M.A., 2009, Improving performance via mini-applications
[12]  
Panda D.K., 2013, WORKSH SUST SOFTW SC
[13]  
Petrini F., 2001, HOT INTERCONNECTS
[14]  
Underwood KD, 2004, PROC INT CONF PARAL, P152
[15]   An Efficient MPI Message Queue Mechanism for Large-scale Jobs [J].
Zounmevo, Judicael A. ;
Afsahi, Ahmad .
PROCEEDINGS OF THE 2012 IEEE 18TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS 2012), 2012, :464-471