By Topic

Using FOM predicting method for scheduling on Chip Multi-Processor

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Gangyong Jia ; Department of Computer Science and Technology, University of Science and Technology of China (USTC), Hefei, 230027, China ; Wei Sheng ; Wenbo Dai ; Xi Li

On a Chip Multi-Processor (CMP) architecture, cache sharing impacts threads non-uniformly, where some threads may be slowed down significantly, while others are not. This may cause severe performance problems such as throughput decreasing, cache thrashing. This paper proposes a new predicting inter-thread cache contention model, FOM (Frequency of Miss), and schedules threads based on the results of FOM on the CMP architecture. The input to our model is the L2 cache misses number of each thread. The output of the model is the extra L2 cache misses for each thread due to cache sharing. We use the output of the model to guide scheduling. We use the multi2sim simulator to compare the throughput of FOA (Frequency of Access) with our FOM, and find our method improving performance up to 13%.

Published in:

Communication Software and Networks (ICCSN), 2011 IEEE 3rd International Conference on

Date of Conference:

27-29 May 2011