Lightweight Provenance Service for High-Performance Computing

被引:11
|
作者
Dai, Dong [1 ]
Chen, Yong [1 ]
Carns, Philip [2 ]
Jenkins, John [2 ]
Ross, Robert [2 ]
机构
[1] Texas Tech Univ, Comp Sci Dept, Lubbock, TX 79409 USA
[2] Argonne Natl Lab, Math & Comp Sci Div, Argonne, IL 60439 USA
来源
2017 26TH INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES (PACT) | 2017年
基金
美国国家科学基金会;
关键词
TIME;
D O I
10.1109/PACT.2017.14
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Provenance describes detailed information about the history of a piece of data, containing the relationships among elements such as users, processes, jobs, and workflows that contribute to the existence of data. Provenance is key to supporting many data management functionalities that are increasingly important in operations such as identifying data sources, parameters, or assumptions behind a given result; auditing data usage; or understanding details about how inputs are transformed into outputs. Despite its importance, however, provenance support is largely underdeveloped in highly parallel architectures and systems. One major challenge is the demanding requirements of providing provenance service in situ. The need to remain lightweight and to be always on often conflicts with the need to be transparent and offer an accurate catalog of details regarding the applications and systems. To tackle this challenge, we introduce a lightweight provenance service, called LPS, for high-performance computing (HPC) systems. LPS leverages a kernel instrument mechanism to achieve transparency and introduces representative execution and flexible granularity to capture comprehensive provenance with controllable overhead. Extensive evaluations and use cases have confirmed its efficiency and usability. We believe that LPS can be integrated into current and future HPC systems to support a variety of data management needs.
引用
收藏
页码:117 / 129
页数:13
相关论文
共 50 条
  • [1] Enabling High-Performance Computing as a Service
    AbdelBaky, Moustafa
    Parashar, Manish
    Kim, Hyunjoo
    Jordan, Kirk E.
    Sachdeva, Vipin
    Sexton, James
    Jamjoom, Hani
    Shae, Zon-Yin
    Pencheva, Gergina
    Tavakoli, Reza
    Wheeler, Mary F.
    COMPUTER, 2012, 45 (10) : 72 - 80
  • [2] Analysis of a lightweight Transport Protocol for high-performance computing
    Dantas, MAR
    Lima, MVGR
    Rodrigues, MRA
    HIGH PERFORMANCE COMPUTING SYSTEMS AND APPLICATIONS, 2002, 657 : 291 - 299
  • [3] High-performance computing service for bioinformatics and data science
    Courneya, Jean-Paul
    Mayo, Alexa
    JOURNAL OF THE MEDICAL LIBRARY ASSOCIATION, 2018, 106 (04) : 494 - 495
  • [4] Teaching high-performance service in a cluster computing course
    Lopez, Pedro
    Baydal, Elvira
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2018, 117 : 138 - 147
  • [5] Application service providing for distributed high-performance computing
    Lee, CK
    Hochberger, C
    Tavangarian, D
    HIGH PERFORMANCE COMPUTING SYSTEMS AND APPLICATIONS, 2003, 727 : 119 - 128
  • [6] A Lightweight Task Graph Scheduler for Distributed High-Performance Scientific Computing
    Weinbub, Josef
    Rupp, Karl
    Selberherr, Siegfried
    APPLIED PARALLEL AND SCIENTIFIC COMPUTING (PARA 2012), 2013, 7782 : 563 - 566
  • [7] USING SERVICE LEVEL AGREEMENTS IN A HIGH-PERFORMANCE COMPUTING ENVIRONMENT
    Kuebert, Roland
    Wesner, Stefan
    SCALABLE COMPUTING-PRACTICE AND EXPERIENCE, 2011, 12 (02): : 164 - 177
  • [8] BLAST: broadband lightweight ATM secure transport for high-performance distributed computing
    Dowd, PW
    Carrozzi, TM
    Pellegrino, FA
    Chen, AX
    Jaeger, R
    Srinidhi, S
    COMPUTER COMMUNICATIONS, 1998, 21 (12) : 1040 - 1057
  • [9] High-Performance Computing
    Bungartz, Hans-Joachim
    IT-INFORMATION TECHNOLOGY, 2013, 55 (03): : 83 - 85