Relative Performance Prediction using Few-Shot Learning

被引:0
作者
Dey, Arunavo [1 ]
Dhakal, Aakash [1 ]
Islam, Tanzima Z. [1 ]
Yeom, Jae-Seung [2 ]
Patki, Tapasya [2 ]
Nichols, Daniel [3 ]
Movsesyan, Alexander [3 ]
Bhatele, Abhinav [3 ]
机构
[1] Texas State Univ, San Marcos, TX 78666 USA
[2] Lawrence Livermore Natl Lab, Livermore, CA USA
[3] Univ Maryland, College Pk, MD USA
来源
2024 IEEE 48TH ANNUAL COMPUTERS, SOFTWARE, AND APPLICATIONS CONFERENCE, COMPSAC 2024 | 2024年
关键词
Performance Modeling; Machine Learning; Few-shot Learning; Large language models (LLM); Cross-platform performance prediction; FRAMEWORK;
D O I
10.1109/COMPSAC61105.2024.00278
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
High-performance computing system architectures are evolving rapidly, making exhaustive data collection for each architecture to build predictive performance models increasingly impractical. Concurrently, the arrival of new applications daily necessitates efficient performance prediction methods. Traditional data collection can take days or weeks, making it more efficient for scientists to leverage existing models to predict an application's performance on new architectures or use data from one application to predict another on the same architecture. The growing heterogeneity in applications and resources further complicates the exact matches needed for effective knowledge transfer. This work systematically studies various Machine Learning (ML) models to predict the relative performance of new applications on new platforms using existing data. Our findings demonstrate that few-shot learning using a few samples significantly enhances cross-platform knowledge transfer, multi-source models outperform single-source models, and Large Language Models (LLMs)-generated samples can effectively improve knowledge transfer efficacy.
引用
收藏
页码:1764 / 1769
页数:6
相关论文
共 16 条
[1]   Cross-Architecture Performance Prediction (XAPP) Using CPU Code to Predict GPU Performance [J].
Ardalani, Newsha ;
Lestourgeon, Clint ;
Sankaralingam, Karthikeyan ;
Zhu, Xiaojin .
PROCEEDINGS OF THE 48TH ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO-48), 2015, :725-737
[2]  
Borisov V., 2023, Language models are realistic tabular data generators
[3]   A performance prediction framework for scientific applications [J].
Carrington, L ;
Snavely, A ;
Wolter, N .
FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2006, 22 (03) :336-346
[4]   FASE: A framework for scalable performance prediction of HPC systems and applications [J].
Grobelny, Eric ;
Bueno, David ;
Troxel, Ian ;
George, Alan D. ;
Vetter, Jeffrey S. .
SIMULATION-TRANSACTIONS OF THE SOCIETY FOR MODELING AND SIMULATION INTERNATIONAL, 2007, 83 (10) :721-745
[5]  
HuggingFace, 2019, DistilGPT2
[6]   Towards HPC I/O Performance Prediction Through Large-scale Log Analysis [J].
Kim, Sunggon ;
Sim, Alex ;
Wu, Kesheng ;
Byna, Suren ;
Son, Yongseok ;
Eom, Hyeonsang .
PROCEEDINGS OF THE 29TH INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE PARALLEL AND DISTRIBUTED COMPUTING, HPDC 2020, 2020, :77-88
[7]   Methods of Inference and Learning for Performance Modeling of Parallel Applications [J].
Lee, Benjamin C. ;
Brooks, David M. ;
de Supinski, Bronis R. ;
Schulz, Martin ;
Singh, Karan ;
McKee, Sally A. .
PROCEEDINGS OF THE 2007 ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING PPOPP'07, 2007, :249-258
[8]   Performance Modeling under Resource Constraints Using Deep Transfer Learning [J].
Marathe, Aniruddha ;
Anirudh, Rushil ;
Jain, Nikhil ;
Bhatele, Abhinav ;
Thiagarajan, Jayaraman ;
Kailkhura, Bhavya ;
Yeom, Jae-Seung ;
Rountree, Barry ;
Gamblin, Todd .
SC'17: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2017,
[9]  
Nichols D., Predicting crossarchitecture performance of parallel programs
[10]   PACE - A toolset for the performance prediction of parallel and distributed systems [J].
Nudd, GR ;
Kerbyson, DJ ;
Papaefstathiou, E ;
Perry, SC ;
Harper, JS ;
Wilcox, DV .
INTERNATIONAL JOURNAL OF HIGH PERFORMANCE COMPUTING APPLICATIONS, 2000, 14 (03) :228-251