Probabilistic relational PCA (PRPCA) can learn a projection matrix to perform dimensionality reduction for relational data. However, the results learned by PRPCA lack interpretabi...
A large family of algorithms for dimensionality reduction end with solving a Trace Ratio problem in the form of arg maxW Tr(WT SpW)/Tr(WT SlW)1 , which is generally transformed in...
7 Linear discriminant analysis (LDA) is a dimension reduction method which finds an optimal linear transformation that maximizes the class separability. However, in undersampled p...
Parameterized Appearance Models (PAMs) (e.g. eigentracking, active appearance models, morphable models) use Principal Component Analysis (PCA) to model the shape and appearance of...
Approximating non-linear kernels using feature maps has gained a lot of interest in recent years due to applications in reducing training and testing times of SVM classifiers and...