By Topic

Prefetching and caching for query scheduling in a special class of distributed applications

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

2 Author(s)
A. Sinha ; Parallel & Distributed Syst. Lab., Texas Univ., Austin, TX, USA ; C. Chase

We analyze the scheduling aspects of database queries submitted to an abstract model of a very large distributed system. The essential elements of this model are: (a) a finite number of identical processing nodes with limited storage capacity, (b) a finite number of queries to be serviced, (c) a very large read-only data set that is shared by all queries and (d) a fixed inter-node communication latency. This framework models an important class of applications that use distributed processing of very large data sets. Examples of these applications exist in the very large database and multimedia problem domains. To meet the objective of minimizing flow time of queries while exploiting inter-query locality, various heuristics are proposed and evaluated through extensive simulation

Published in:

Parallel Processing, 1996. Vol.3. Software., Proceedings of the 1996 International Conference on  (Volume:3 )

Date of Conference:

12-16 Aug 1996