By Topic

Isoefficiency Maps for Divisible Computations

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Maciej Drozdowski ; Poznan University of Technology, Poznan ; Lukasz Wielebski

In this paper, we propose a new technique of presenting performance relationships in parallel processing. Performance of parallel processing is a hard matter with many counterintuitive phenomena. It is relatively easy to obtain some numerical indicators of the performance using various performance models. However, it is far more difficult to comprehend the nature of the analyzed problem. To facilitate understanding the performance relationships, we propose a new visualization technique based on the concept of isoefficiency. In this paper, isoefficiency is represented as a relation on points in the space of system parameters for which efficiency of parallel processing is equal. We visualize this relation on two-dimensional maps analogously to isobars and isotherms on weather maps. This concept is applied to depict the performance relationships in two standard performance laws: Amdahl's speedup law and Gustafson's speedup law. Then, we use isoefficiency maps to analyze the performance relationships in divisible load processing. Divisible load model conforms with data-parallel computations in an environment with communication delays. The results we obtain give interesting insights into relationships existing in parallel processing.

Published in:

IEEE Transactions on Parallel and Distributed Systems  (Volume:21 ,  Issue: 6 )