By Topic

Efficient Parallel Algorithm for Nonlinear Dimensionality Reduction on GPU

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$33 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

4 Author(s)
Tsung Tai Yeh ; Inst. of Inf. Sci., Acad. Sinica Taipei, Taipei, Taiwan ; Tseng-Yi Chen ; Yen-Chiu Chen ; Wei-Kuan Shih

Advances in nonlinear dimensionality reduction provide a way to understand and visualize the underlying structure of complex data sets. The performance of large-scale nonlinear dimensionality reduction is of key importance in data mining, machine learning, and data analysis. In this paper, we concentrate on improving the performance of nonlinear dimensionality reduction using large-scale data sets on the GPU. In particular, we focus on solving problems including k nearest neighbor (KNN) search and sparse spectral decomposition for large-scale data, and propose an efficient framework for Local Linear Embedding (LLE). We implement a k-d tree based KNN algorithm and Krylov subspace method on the GPU to accelerate the nonlinear dimensionality reduction for large-scale data. Our results enable GPU-based k-d tree LLE processes of up to about 30-60 X faster compared to the brute force KNN LLE model on the CPU. Overall, our methods save O (n2-6n-2k-3) memory space.

Published in:

Granular Computing (GrC), 2010 IEEE International Conference on

Date of Conference:

14-16 Aug. 2010