Lightweight Provenance Service for High-Performance Computing

被引:11
|
作者
Dai, Dong [1 ]
Chen, Yong [1 ]
Carns, Philip [2 ]
Jenkins, John [2 ]
Ross, Robert [2 ]
机构
[1] Texas Tech Univ, Comp Sci Dept, Lubbock, TX 79409 USA
[2] Argonne Natl Lab, Math & Comp Sci Div, Argonne, IL 60439 USA
来源
2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT) | 2017年
基金
美国国家科学基金会;
关键词
TIME;
D O I
10.1109/PACT.2017.14
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Provenance describes detailed information about the history of a piece of data, containing the relationships among elements such as users, processes, jobs, and workflows that contribute to the existence of data. Provenance is key to supporting many data management functionalities that are increasingly important in operations such as identifying data sources, parameters, or assumptions behind a given result; auditing data usage; or understanding details about how inputs are transformed into outputs. Despite its importance, however, provenance support is largely underdeveloped in highly parallel architectures and systems. One major challenge is the demanding requirements of providing provenance service in situ. The need to remain lightweight and to be always on often conflicts with the need to be transparent and offer an accurate catalog of details regarding the applications and systems. To tackle this challenge, we introduce a lightweight provenance service, called LPS, for high-performance computing (HPC) systems. LPS leverages a kernel instrument mechanism to achieve transparency and introduces representative execution and flexible granularity to capture comprehensive provenance with controllable overhead. Extensive evaluations and use cases have confirmed its efficiency and usability. We believe that LPS can be integrated into current and future HPC systems to support a variety of data management needs.
引用
收藏
页码:117 / 129
页数:13
相关论文
共 50 条
  • [31] High-Performance Computing for Defense
    Davis, Larry P.
    Henry, Cray J.
    Campbell, Roy L., Jr.
    Ward, William A., Jr.
    COMPUTING IN SCIENCE & ENGINEERING, 2007, 9 (06) : 35 - 44
  • [32] The marketplace of high-performance computing
    Strohmaier, E
    Dongarra, JJ
    Meuer, HW
    Simon, HD
    PARALLEL COMPUTING, 1999, 25 (13-14) : 1517 - 1544
  • [33] Optical high-performance computing
    Fisk University, Nashville, TN, United States
    不详
    不详
    Journal of the Optical Society of America A: Optics and Image Science, and Vision, 2008, 25 (09):
  • [34] Productivity in high-performance computing
    Sterling, Thomas
    Dekate, Chirag
    ADVANCES IN COMPUTERS, VOL 72: HIGH PERFORMANCE COMPUTING, 2008, 72 : 101 - 134
  • [35] High-performance computing - An overview
    Marksteiner, P
    COMPUTER PHYSICS COMMUNICATIONS, 1996, 97 (1-2) : 16 - 35
  • [36] HIGH-PERFORMANCE DISTRIBUTED COMPUTING
    RAGHAVENDRA, CS
    CONCURRENCY-PRACTICE AND EXPERIENCE, 1994, 6 (04): : 231 - 233
  • [37] High-Performance Computing with TeraStat
    Bompiani, Edoardo
    Petrillo, Umberto Ferraro
    Lasinio, Giovanna Jona
    Palini, Francesco
    2020 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS (DASC/PICOM/CBDCOM/CYBERSCITECH), 2020, : 499 - 506
  • [38] High-Performance Computing for All
    Wince-Smith, Deborah
    ISSUES IN SCIENCE AND TECHNOLOGY, 2009, 25 (04) : 67 - 70
  • [39] The high-performance computing continuum
    Karin, S
    Graham, S
    COMMUNICATIONS OF THE ACM, 1998, 41 (11) : 32 - 35
  • [40] HIGH-PERFORMANCE COMPUTING IN EUROPE
    EMMEN, A
    HARMS, U
    SUPERCOMPUTER, 1995, 11 (2-3): : 46 - 60