By Topic

Optimizing the parSOM neural network implementation for data mining with distributed memory systems and cluster computing

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Tomsich, P. ; Inst. of Software Technol., Vienna Univ. of Technol., Austria ; Rauber, A. ; Merkl, D.

The self-organizing map is a prominent unsupervised neural network model which lends itself to the analysis of high-dimensional input data and data mining applications. However, the high execution times required to train the map limit its application in many high-performance data analysis application domains. We discuss the parSOM implementation, a software-based parallel implementation of the self-organizing map, and its optimization for the analysis of high-dimensional input data using distributed memory systems and clusters. The original parSOM algorithm scales very well in a parallel execution environment with low communication latencies and exploits parallelism to cope with memory latencies. However it suffers from poor scalability on distributed memory computers. We present optimizations to further decouple the subprocesses, simplify the communication model and improve the portability of the system

Published in:

Database and Expert Systems Applications, 2000. Proceedings. 11th International Workshop on

Date of Conference: