Compiler directed parallelization of loops in scale for shared-memory multiprocessors

被引:0
|
作者
Johnson, GS [1 ]
Sethumadhavan, S
机构
[1] Univ Texas, Dept Comp Sci, Austin, TX 78712 USA
[2] Univ Texas, Texas Adv Comp Ctr, Austin, TX 78712 USA
来源
COMPUTATIONAL SCIENCE - ICCS 2003, PT III, PROCEEDINGS | 2003年 / 2659卷
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Effective utilization of symmetric shared-memory multiprocessors (SMPs) is predicated on the development of efficient parallel code. Unfortunately, efficient parallelism is not always easy for the programmer to identify. Worse, exploiting such parallelism may directly conflict with optimizations affecting per-processor utilization (i.e. loop reordering to improve data locality). Here, we present our experience with a loop-level parallel compiler optimization for SMPs proposed by McKinley [6]. The algorithm uses dependence analysis and a simple model of the target machine, to transform nested loops. The goal of the approach is to promote efficient execution of parallel loops by exposing sources of large-grain parallel work while maintaining per-processor locality. We implement the optimization within the Scale compiler framework, and analyze the performance of multiprocessor code produced for three microbenchmarks.
引用
收藏
页码:946 / 955
页数:10
相关论文
共 50 条