By Topic

Matrix-based Kernel Principal Component analysis for large-scale data set

Sign In

Cookies must be enabled to login.After enabling cookies , please use refresh or reload or ctrl+f5 on the browser for the login options.

Formats Non-Member Member
$31 $13
Learn how you can qualify for the best price for this item!
Become an IEEE Member or Subscribe to
IEEE Xplore for exclusive pricing!
close button

puzzle piece

IEEE membership options for an individual and IEEE Xplore subscriptions for an organization offer the most affordable access to essential journal articles, conference papers, standards, eBooks, and eLearning courses.

Learn more about:

IEEE membership

IEEE Xplore subscriptions

3 Author(s)
Weiya Shi ; Sch. of Comput. Sci., Fudan Univ., Shanghai, China ; Yue-Fei Guo ; Xiangyang Xue

Kernel Principal Component Analysis (KPCA) is a nonlinear feature extraction approach, which generally needs to eigen-decompose the kernel matrix. But the size of kernel matrix scales with the number of data points, it is infeasible to store and compute the kernel matrix when faced with the large-scale data set. To overcome computational and storage problem for large-scale data set, a new framework, Matrixbased Kernel Principal Component Analysis (M-KPCA), is proposed. By dividing the large scale data set into small subsets, we could treat the autocorrelation matrix of each subset as the special computational unit. A novel polynomial-matrix kernel function is adopted to compute the similarity between the data matrices in place of vectors. It is also proved that the polynomial kernel is the extreme case of the polynomial-matrix one. The proposed M-KPCA can greatly reduce the size of kernel matrix, which makes its computation possible. The effectiveness is demonstrated by the experimental results on the artificial and real data set.

Published in:

Neural Networks, 2009. IJCNN 2009. International Joint Conference on

Date of Conference:

14-19 June 2009