📄 readme.txt
字号:
Matlab Toolbox for Dimensionality Reduction (v0.4b)===================================================Information-------------------------Author: Laurens van der MaatenAffiliation: MICC, Maastricht University, The NetherlandsContact: l.vandermaaten@micc.unimaas.nlRelease date: January 12, 2008Version: 0.4bInstallation------------Copy the drtoolbox/ folder into the $MATLAB_DIR/toolbox directory (where $MATLAB_DIR indicates your Matlab installation directory). Start Matlab and select Set path... from the File menu. Click the Add with subfolders... button, select the folder $MATLAB_DIR/toolbox/drtoolbox in the file dialog, and press Open. Subsequently, press the Save button in order to save your changes to the Matlab search path. The toolbox is now installed. Some of the functions in the toolbox use MEX-files. Precompiled versions of these MEX-files are distributed with this release, but the compiled version for your platform might be missing. In order to compile all MEX-files, type cd([matlabroot '/toolbox/drtoolbox']) in your Matlab prompt, and execute the function MEXALL. Features-------------------------This Matlab toolbox implements 29 techniques for dimensionality reduction. These techniques are all available through the COMPUTE_MAPPING function. The following techniques are available: - Principal Component Analysis ('PCA') - Linear Discriminant Analysis ('LDA') - Multidimensional scaling ('MDS') - Simple PCA ('SimplePCA') - Probabilistic PCA ('ProbPCA') - Factor analysis ('FactorAnalysis') - Isomap ('Isomap') - Landmark Isomap ('LandmarkIsomap') - Locally Linear Embedding ('LLE') - Laplacian Eigenmaps ('Laplacian') - Hessian LLE ('HessianLLE') - Local Tangent Space Alignment ('LTSA') - Diffusion maps ('DiffusionMaps') - Kernel PCA ('KernelPCA') - Generalized Discriminant Analysis ('KernelLDA') - Stochastic Neighbor Embedding ('SNE') - Neighborhood Preserving Embedding ('NPE') - Linearity Preserving Projection ('LPP') - Stochastic Proximity Embedding ('SPE') - Linear Local Tangent Space Alignment ('LLTSA') - Conformal Eigenmaps ('CCA', implemented as an extension of LLE) - Maximum Variance Unfolding ('MVU', implemented as an extension of LLE) - Landmark Maximum Variance Unfolding ('LandmarkMVU') - Fast Maximum Variance Unfolding ('FastMVU') - Locally Linear Coordination ('LLC') - Manifold charting ('ManifoldChart') - Coordinated Factor Analysis ('CFA') - Autoencoders using stack-of-RBMs pretraining ('AutoEncoderRBM') - Autoencoders using evolutionary optimization ('AutoEncoderEA')Furthermore, the toolbox contains 6 techniques for intrinsic dimensionality estimation. These techniques are available through the function INTRINSIC_DIM. The following techniques are available: - Eigenvalue-based estimation ('EigValue') - Maximum Likelihood Estimator ('MLE') - Estimator based on correlation dimension ('CorrDim') - Estimator based on nearest neighbor evaluation ('NearNb') - Estimator based on packing numbers ('PackingNumbers') - Estimator based on geodesic minimum spanning tree ('GMST')In addition to these techniques, the toolbox contains functions for prewhitening of data (the function PREWHITEN), exact and estimate out-of-sample extension (the functions OUT_OF_SAMPLE and OUT_OF_SAMPLE_EST), and a function that generates toy datasets (the function GENERATE_DATA).Usage-------------------------Basically, you only need one function: mappedX = compute_mapping(X, technique, no_dims);Try executing the following code: [X, labels] = generate_data('helix', 2000); figure, scatter3(X(:,1), X(:,2), X(:,3), 5, labels); title('Original dataset'), drawnow no_dims = round(intrinsic_dim(X, 'MLE')); disp(['MLE estimate of intrinsic dimensionality: ' num2str(no_dims)]); mappedX = compute_mapping(X, 'Laplacian', no_dims, 7); figure, scatter(mappedX(:,1), mappedX(:,2), 5, labels); title('Result of dimensionality reduction'), drawnowIt will create a helix dataset, estimate the intrinsic dimensionality of the dataset, run Laplacian Eigenmaps on the dataset, and plot the results. All functions in the toolbox can work both on data matrices as on PRTools datasets (http://prtools.org). For more information on the options for dimensionality reduction, type HELP COMPUTE_MAPPING in your Matlab prompt. Information on the intrinsic dimensionality estimators can be obtained by typing the HELP INTRINSIC_DIM.Other functions that are useful are the GENERATE_DATA function and the OUT_OF_SAMPLE and OUT_OF_SAMPLE_EST functions. The GENERATE_DATA function provides you with a number of artificial datasets to test the techniques. The OUT_OF_SAMPLE function allows for out-of-sample extension for the techniques PCA, LDA, LPP, NPE, LLTSA, Kernel PCA, and autoencoders. The OUT_OF_SAMPLE_EST function allows you to perform an out-of-sample extension using an estimation technique, that is generally applicable.Pitfalls-------------------------When you run certain code, you might receive an error that a certain file is missing. This is because in some parts of the code, MEX-functions are used. I provide a number of precompiled versions of these MEX-functions in the toolbox. However, the MEX-file for your platform might be missing. To fix this, type in your Matlab: mexallNow you have compiled versions of the MEX-files as well. This fix also solves slow execution of the shortest path computations in Isomap.If you encounter an error considering CSDP while running the FastMVU-algorithm, the binary of CSDP for your platform is missing. If so, please obtain a binary distribution of CSDP from https://projects.coin-or.org/Csdp/ and place it in the drtoolbox/techniques directory. Make sure it has the right name for your platform (csdp.exe for Windows, csdpmac for Mac OS X (PowerPC), csdpmaci for Mac OS X (Intel), and csdplinux for Linux).Many methods for dimensionality reduction perform spectral analyses of sparse matrices. You might think that eigenanalysis is a well-studied problem that can easily be solved. However, eigenanalysis of large matrices turns out to be tedious. The toolbox allows you to use two different methods for eigenanalysis: - The original Matlab functions (based on Arnoldi methods) - The JDQR functions (based on Jacobi-Davidson methods)For problems up to 10,000 datapoints, we recommend using the 'Matlab' setting. For larger problems, switching to 'JDQR' is often worth trying.Papers-------------------------For more information on the implemented techniques and for a theoretical and empirical comparison, please have a look at the following papers:
⌨️ 快捷键说明
复制代码
Ctrl + C
搜索代码
Ctrl + F
全屏模式
F11
切换主题
Ctrl + Shift + D
显示快捷键
?
增大字号
Ctrl + =
减小字号
Ctrl + -