i tried this code for my data set of size 400x2048.2048 dimensions and 400 observations.the results were not code.won't it work for large data?what should i do for best solution.i have to reduce my data to 220 dimensions.any help please..

I think there are some mistake in this implementation, the last step the feature vector feature dimension reduction procedure is incorrect, since you can not do it in this way. If you do it in this way, how can you tell the difference between PCA and KPCA. we should do it by using inner product form.

I am unable to run the algorithm for large data set. I am working on hyperspectral image of size 191x370856. What can I do and what is the best solution.

Sorry Enrique I don't understand your second point. Surely it doesn't matter whether you normalize just the non-zero eigenvalues or all of the eigenvalues since the non-zero eigenvalues won't change the projections? Have I missed something? Thanks.

I used your algorithm and works fine, i have taken the entire data that is 5x3000 and after simulation i got new data, can i use this new data for Support vector regression.

Thank you for a great submission Ambarish, helped a lot to figure out Schölkopf's "Nonlinear Component Analysis as a Kernel Eigenvalue
Problem".

And thank for dear comments Enrique, I have thought alot about 1) and 2) as well but was afraid I was the one mistaken as you often are when learning new stuff.

Sorry about my last two blank comments. Mouse double-click errors.

Very nice code! However I have a couple of subtle comments:

1) When you center the Kernel matrix be sure to divide the "ones" matrix by the number of samples i.e.
Line 43: one_mat = ones(size(K));
should read
Line 43: one_mat = ones(size(K))./size(data_in,2);

2) Eigenvector normalization implies dividing each of the columns of the eigenvector matrix by the sqrt of its corresponding eigenvalue. Do so by substituting:
Line 61: eigvec(:,col) = eigvec(:,col)./(sqrt(eig_val(col,col)));
to read
Line 61: eigvec(:,col) = eigvec(:,col)./(sqrt(eigval(col,col)));

3) Also, if you must do eigenvalue sorting, be careful to use only the diagonal of the "Lambda" matrix and not the whole matrix. Use:

Line 63: [dummy, index] = sort(diag(eigval),'descend');
instead of
Line 63: [dummy, index] = sort(eig_val,'descend');

You can also see "Learning with Kernels" by B. Scholkopf and A. Smola, Section 14.2 (particularly eqs 14.14 and 14.17). Hope this is helpful.

I use Matlab 2008 (7.6) and there is error at this line:
[~, index] = sort(eig_val,'descend');
Expression or statement is incorrect--possibly unbalanced (, {, or [.