Kernel Multivariate Analysis Framework for Supervised Subspace Learning Articles uri icon

publication date

  • July 2013

start page

  • 16

end page

  • 29

issue

  • 4

volume

  • 30

International Standard Serial Number (ISSN)

  • 1053-5888

Electronic International Standard Serial Number (EISSN)

  • 1558-0792

abstract

  • Feature extraction and dimensionality reduction are important tasks in many fields of science dealing with signal processing and analysis. The relevance of these techniques is increasing as current sensory devices are developed with ever higher resolution, and problems involving multimodal data sources become more common. A plethora of feature extraction methods are available in the literature collectively grouped under the field of multivariate analysis (MVA). This article provides a uniform treatment of several methods: principal component analysis (PCA), partial least squares (PLS), canonical correlation analysis (CCA), and orthonormalized PLS (OPLS), as well as their nonlinear extensions derived by means of the theory of reproducing kernel Hilbert spaces (RKHSs). We also review their connections to other methods for classification and statistical dependence estimation and introduce some recent developments to deal with the extreme cases of large-scale and low-sized problems. To illustrate the wide applicability of these methods in both classification and regression problems, we analyze their performance in a benchmark of publicly available data sets and pay special attention to specific real applications involving audio processing for music genre prediction and hyperspectral satellite image processing for Earth and climate monitoring.

keywords

  • audio signal processing; feature extraction; hyperspectral imaging; image processing; learning (artificial intelligence); least squares approximations; music; principal component analysis; regression analysis; remote sensing; signal classification