Cart (Loading....) | Create Account
Close category search window
 

Trident: technology-scalable architecture for data parallel applications

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
Sedukhin, S.G. ; Graduate Sch. of Comput. Sci. & Eng., Univ. of Aizu, Fukushima, Japan ; Soliman, M.I.

Within the current decade, process technology is promising more than one billion transistors on a single die, operating at frequency more than 10 GHz. We proposed the Trident processor, which uses multi-level ISA to express data parallelism to hardware. Trident is scalable because its architecture is regular, which can be widely replicated to efficiently harness the available transistor budget. Besides, it is based on local communication, which is very suitable for a high operating frequency of the future VLSI technology. This paper discusses the Trident processor architecture and evaluates its performance on the Basic Linear Algebra Subprograms (BLAS), which are widely used in many data parallel applications. The TFLOPS rate on infinite-size problems (R), which is primarily a characteristic of the computer technology, and the problem size needed to reach one-half of R (N12/), which is a measure of the amount of parallelism in a computer architecture, are used to evaluate the performance of the Trident processor on BLAS. On 128 parallel Trident lanes and 10 GHz operating frequency, which are possible in the billion-transistor era, R of dot product, matrix-vector, and matrix-matrix multiplications are 1.1, 1.8, and 2.5 TFLOPS, respectively. Besides, N12/ increases when switching from low level to high level of BLAS.

Published in:

Parallel and Distributed Processing Symposium, 2003. Proceedings. International

Date of Conference:

22-26 April 2003

Need Help?


IEEE Advancing Technology for Humanity About IEEE Xplore | Contact | Help | Terms of Use | Nondiscrimination Policy | Site Map | Privacy & Opting Out of Cookies

A not-for-profit organization, IEEE is the world's largest professional association for the advancement of technology.
© Copyright 2014 IEEE - All rights reserved. Use of this web site signifies your agreement to the terms and conditions.