Computational Intelligence - November 2016 - 25

Abstract

Multivariate Analysis (MVA) comprises a family of well-known methods for feature
extraction which exploit correlations among input variables representing the data.
One important property that is enjoyed by most such methods is uncorrelation
among the extracted features. Recently, regularized versions of MVA methods have
appeared in the literature, mainly with the goal to gain interpretability of the solution. In these cases, the solutions can no longer be obtained in a closed manner, and
more complex optimization methods that rely on the iteration of two steps are frequently used. This paper recurs to an alternative approach to solve efficiently this
iterative problem. The main novelty of this approach lies in preserving several properties of the original methods, most notably the uncorrelation of the extracted features. Under this framework, we propose a novel method that takes advantage of the
, 2, 1 norm to perform variable selection during the feature extraction process.
Experimental results over different problems corroborate the advantages of the proposed formulation in comparison to state of the art formulations.

direction, while a supervised method like OPLS (right
subplot) successfully identifies the most discriminative
information. Although this toy example is based on a
classification task, the same advantages of supervised
MVA over standard PCA are encountered in regression
tasks-see [7] for a detailed theoretical and experimental review of these methods.
The simplicity of these methods, as well as the availability of highly-optimized libraries for solving the linear algebra problems they involve, justifies the extensive use of
MVA in many application fields, such as biomedical
engineering [8], [9], remote sensing [10], [11], or chemometrics [12], among many others (see also [7] for
a more detailed review of application-oriented
research in the field).
An important property of PCA, OPLS, and
CCA is that they lead to uncorrelated variables, so that the feature extraction process
provides additional advantages:
❏ The relevance of each extracted feature is
directly given by the magnitude of its associated eigenvalue, which simplifies the
selection of a reduced subset of features,
if necessary.
❏ Subsequent learning tasks are simplified, more
notably, when the covariance matrix inversion is
required. This is the case of least-square based
problems, such as Ridge Regression or lasso (least
absolute shrinkage and selection operator) [13].
Standard versions of MVA methods implement just a feature
extraction process, in the sense that all original variables are used
to build the new features. However, over the last few years there

Corresponding Author: Sergio Muñoz-Romero (e-mail: sergio.munoz@urjc.es).

have been many significant contributions to this field that have
focused on gaining interpretability of the extracted features by
incorporating sparsity-inducing norms, such as the , 1 and , 2, 1
norms [14], as a penalty term in the minimization problem.
When these regularization terms are included, the projection
vectors are favored to include zeros in some of their components, making it easier to understand the process to build the
new features and thus gaining in interpretability. In fact, the , 2, 1
rewards solutions that perform a real variable selection process,
in the sense that some of the original variables are excluded
from all projection vectors at once. In other words, only a subset
of the original variables are used to build the new features.
Some of the most significant contributions in this direction
are sparse PCA [15], sparse OPLS [8], group-lasso penalized
OPLS (also known as Sparse Reduced Rank Regression,
SRRR) [16], and , 2, 1-regularized CCA (or L21SDA) [17]. All
these approaches are based on an iterative process which combines the optimization of two coupled least-squares problems,
one of them subject to a minimization constraint. Since the
inspiring work [15], this constrained least-squares minimization
has been typically treated as an orthogonal Procrustes problem
[18], an approach that can still be considered mainstream (see,
e.g., the very recent works [19], [20]).
A first objective of this paper is to highlight and make the
computational intelligence community aware of some limitations
derived from the use of orthogonal Procrustes in the context of
regularized MVA methods. As explained in [21], these methods
1) do not converge to their associated non-regularized MVA
solutions when the penalty term is removed,
2) are highly dependent on initialization, and may even fail
to progress towards a solution,
3) do not in general obtain uncorrelated features.
As solution to these problems, [21] proposes an alternative optimization procedure avoiding the use of the Procrustes solution. In
this paper, we will briefly review the framework presented in [21]

novEmbEr 2016 | IEEE ComputatIonal IntEllIgEnCE magazInE

25



Table of Contents for the Digital Edition of Computational Intelligence - November 2016

Computational Intelligence - November 2016 - Cover1
Computational Intelligence - November 2016 - Cover2
Computational Intelligence - November 2016 - 1
Computational Intelligence - November 2016 - 2
Computational Intelligence - November 2016 - 3
Computational Intelligence - November 2016 - 4
Computational Intelligence - November 2016 - 5
Computational Intelligence - November 2016 - 6
Computational Intelligence - November 2016 - 7
Computational Intelligence - November 2016 - 8
Computational Intelligence - November 2016 - 9
Computational Intelligence - November 2016 - 10
Computational Intelligence - November 2016 - 11
Computational Intelligence - November 2016 - 12
Computational Intelligence - November 2016 - 13
Computational Intelligence - November 2016 - 14
Computational Intelligence - November 2016 - 15
Computational Intelligence - November 2016 - 16
Computational Intelligence - November 2016 - 17
Computational Intelligence - November 2016 - 18
Computational Intelligence - November 2016 - 19
Computational Intelligence - November 2016 - 20
Computational Intelligence - November 2016 - 21
Computational Intelligence - November 2016 - 22
Computational Intelligence - November 2016 - 23
Computational Intelligence - November 2016 - 24
Computational Intelligence - November 2016 - 25
Computational Intelligence - November 2016 - 26
Computational Intelligence - November 2016 - 27
Computational Intelligence - November 2016 - 28
Computational Intelligence - November 2016 - 29
Computational Intelligence - November 2016 - 30
Computational Intelligence - November 2016 - 31
Computational Intelligence - November 2016 - 32
Computational Intelligence - November 2016 - 33
Computational Intelligence - November 2016 - 34
Computational Intelligence - November 2016 - 35
Computational Intelligence - November 2016 - 36
Computational Intelligence - November 2016 - 37
Computational Intelligence - November 2016 - 38
Computational Intelligence - November 2016 - 39
Computational Intelligence - November 2016 - 40
Computational Intelligence - November 2016 - 41
Computational Intelligence - November 2016 - 42
Computational Intelligence - November 2016 - 43
Computational Intelligence - November 2016 - 44
Computational Intelligence - November 2016 - 45
Computational Intelligence - November 2016 - 46
Computational Intelligence - November 2016 - 47
Computational Intelligence - November 2016 - 48
Computational Intelligence - November 2016 - 49
Computational Intelligence - November 2016 - 50
Computational Intelligence - November 2016 - 51
Computational Intelligence - November 2016 - 52
Computational Intelligence - November 2016 - 53
Computational Intelligence - November 2016 - 54
Computational Intelligence - November 2016 - 55
Computational Intelligence - November 2016 - 56
Computational Intelligence - November 2016 - 57
Computational Intelligence - November 2016 - 58
Computational Intelligence - November 2016 - 59
Computational Intelligence - November 2016 - 60
Computational Intelligence - November 2016 - 61
Computational Intelligence - November 2016 - 62
Computational Intelligence - November 2016 - 63
Computational Intelligence - November 2016 - 64
Computational Intelligence - November 2016 - 65
Computational Intelligence - November 2016 - 66
Computational Intelligence - November 2016 - 67
Computational Intelligence - November 2016 - 68
Computational Intelligence - November 2016 - 69
Computational Intelligence - November 2016 - 70
Computational Intelligence - November 2016 - 71
Computational Intelligence - November 2016 - 72
Computational Intelligence - November 2016 - Cover3
Computational Intelligence - November 2016 - Cover4
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202311
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202308
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202305
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202302
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202211
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202208
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202205
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202202
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202111
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202108
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202105
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202102
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202011
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202008
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202005
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_202002
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201911
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201908
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201905
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201902
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201811
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201808
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201805
https://www.nxtbook.com/nxtbooks/ieee/computationalintelligence_201802
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring17
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring16
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring15
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring14
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_summer13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_spring13
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_winter12
https://www.nxtbook.com/nxtbooks/ieee/computational_intelligence_fall12
https://www.nxtbookmedia.com