• No results found

Chemometrics and Intelligent Laboratory Systems

N/A
N/A
Protected

Academic year: 2022

Share "Chemometrics and Intelligent Laboratory Systems"

Copied!
11
0
0

Bezig met laden.... (Bekijk nu de volledige tekst)

Hele tekst

(1)

Clusterwise Parafac to identify heterogeneity in three-way data

Tom F. Wilderjans ⁎ , Eva Ceulemans

1

Methodology of Educational Sciences Research Group, Faculty of Psychology and Educational Sciences, KU Leuven, Andreas Vesaliusstraat 2, Box 3762, 3000 Leuven, Belgium

a b s t r a c t a r t i c l e i n f o

Article history:

Received 24 December 2012

Received in revised form 19 August 2013 Accepted 20 September 2013 Available online 18 October 2013

Keywords:

Parafac Candecomp Clusterwise analysis Sensory profiling data Population heterogeneity Three-way data

Qualitative (and quantitative) differences between elements

In many research areas, the Parafac model is adopted to disclose the underlying structure of three-way three- mode data. In this model, a set of latent variables, called components, that captures the complex interaction between the elements of the three modes is sought. An important assumption of this model is that these components are the same for all elements of the three modes. In many cases, however, it makes sense to assume that the components may differ (i.e., qualitative differences in underlying component structure) across groups of elements of one of the modes. Therefore, in this paper, we present Clusterwise Parafac. In this new model, the elements of one of the three modes are assigned to a limited number of mutually exclusive clusters and, simultaneously, the data within each cluster are modeled with Parafac. As such, elements that belong to the same cluster are assumed to be governed by the same components, whereas elements that are assigned to different clusters have a different underlying component structure. To evaluate the performance of the new Clusterwise Parafac strategy, an extensive simulation study is conducted. Moreover, the strategy is applied to sensory profiling data regarding different cheeses.

© 2013 Elsevier B.V. All rights reserved.

1. Introduction

In many research areas, the Parafac model [1–3] is used to summarize the covariation in three-way three-mode data. One such research domain is sensory profiling, where one often lets a number of panelists rate a set of food samples (e.g., different cheeses) on a number of sensory attributes [4]. In chemometrics, this model is a popular method to explore the structure offluorescence spectroscopy data [see,5–12] or of chromatography results [see, a.o.,13–15]. Still other application areas include mass spectrometry [16,17], second order calibration[18,19], and (on-line) batch process monitoring[20].

In Parafac, components (i.e., latent variables) are sought that capture the complex interaction between the elements of the three modes. In the sensory profiling example, the components represent the most important dimensions for distinguishing between the cheese samples, for example, the taste and the texture. Each element of each mode receives a score on these components. The attribute scores reflect to which extent the attributes measure the underlying dimensions, whereas the sample scores indicate the location of each sample on these dimensions (e.g., the texture of a particular sample). The panelist scores represent the extent to which the panelists rely on a particular dimension when rating the samples (i.e., saliency).

An important assumption of this model is that the components are the same for all elements of the three modes. In many cases, however,

it makes sense to assume that the underlying dimensions may differ across the elements of one of the modes [i.e., qualitative differences in underlying component structure may exist between elements of one mode; for a discussion of qualitative and quantitative differences in three-way data, see 21]. For the sensory profiling example, it is reasonable to assume that the dimensions that are important for capturing the quality of French soft cheeses differ from those used for evaluating other cheese types, such as blue cheeses. To complicate things further, it is often not known beforehand which elements are scored in a similar way and thus have a similar component structure and which are not. In order to account for such qualitative differences between the elements of a particular mode, a data analysis technique is needed that induces which elements are similar in that their ratings can be summarized by means of the same components, and which elements are not.

Therefore, we will introduce a new modeling strategy, called Clusterwise Parafac. In this new model, the elements of one of the three modes are assigned to a limited number of mutually exclusive clusters and, simultaneously, the data within each cluster are modeled with Parafac. Hence, qualitative differences between elements are modeled by assigning elements that have a similar component structure to the same cluster and elements with different structures to different clusters.

The remainder of this paper is organised infive main sections: In Section 2, the Parafac model is recapitulated and the new Clusterwise Parafac model is presented;finally, its relations to existing cluster models for three-way data are discussed. In Section 3, we will discuss an alternating least squares algorithm to estimate the parameters of the new model, along with a model selection procedure. InSection 4, we will present an extensive simulation study to evaluate the performance

⁎ Corresponding author. Tel.: +32 16 32 61 23, +32 16 32 62 01; fax: +32 16 32 62 00.

E-mail addresses:tom.wilderjans@ppw.kuleuven.be(T.F. Wilderjans), eva.ceulemans@ppw.kuleuven.be(E. Ceulemans).

1Tel.: +32 16 32 58 81, +32 16 32 62 01; fax: +32 16 32 62 00.

0169-7439/$– see front matter © 2013 Elsevier B.V. All rights reserved.

http://dx.doi.org/10.1016/j.chemolab.2013.09.010

Contents lists available atScienceDirect

Chemometrics and Intelligent Laboratory Systems

j o u r n a l h o m e p a g e : w w w . e l s e v i e r . c o m / l o c a t e / c h e m o l a b

(2)

of the Clusterwise Parafac algorithm in terms of optimizing the loss function and recovering the underlying parameters. InSection 5, an illustrative application of the new model to sensory profiling data of cheese samples will be discussed. Finally, inSection 6, we will conclude with a general discussion.

2. Model

2.1. The Parafac model for three-way three-mode data

In Parafac[1–3], the I × J × K variable by source by object model array M isfitted to the I × J × K data array X. The model array M is further decomposed into an I × R variable component matrix A, a J × R source component matrix B, and a K×R object component matrix C. In particular, the model matrix Mk(of size I×J) of object k (k=1…K) is decomposed as follows:

Mk¼ ADkBT ð1Þ

where Dkis a R × R diagonal matrix that has the component scores for object k (i.e., the k-th row of C) on its diagonal (k = 1…K), T denotes matrix transposition, and R denotes the number of components. As A and B have no index k, they are assumed to be the same for all objects.

Under mild conditions, the Parafac decomposition is unique, except for a permutation and a scaling of the components[22–26].

2.2. The Clusterwise Parafac model

In order to allow clusters of objects2to be governed by (qualitatively) different components, Clusterwise Parafac partitions the K objects into Q mutually exclusive and non-empty clusters. Objects that belong to the same cluster have the same underlying components, whereas objects that belong to different clusters are assumed to have a different component structure. Specifically, in Clusterwise Parafac the model matrix Mkfor object k (that belongs to cluster q) is decomposed as follows:

Mk¼XQ

q¼1

pkqAqDkBTq

ð2Þ

where Q is the number of clusters, pkqare the entries of the binary partition matrix P (K × Q) which indicates whether object k belongs to cluster q (pkq= 1) or not (pkq= 0), Aq(I × R) is the variable component matrix and Bq(J × R) the source component matrix for cluster q (q = 1…Q). Note that the number of components R is assumed to be equal for all objects, which in some cases may be too restrictive (for a further discussion of this topic, we refer toSection 6).3

From the Clusterwise Parafac decomposition rule in Eq.(2), it can be seen that this model is a generic modeling strategy that encompasses different models as special cases, two of which will be outlined here.

First, when all objects are assigned to the same cluster (i.e., Q equals one), the standard Parafac model is obtained. A second special case is encountered when all objects are assigned to a different cluster (i.e., Q equals K), which implies that each object is modeled by a separate set

of components. In this case, the new model boils down to performing a separate PCA to each I × J data matrix Xk, which consists of the data slice of X that pertains to object k.

Note that a Clusterwise Parafac model with Q clusters and R components can also be conceived as a constrained Parafac model with Q × R components. Specifically, in the latter model, the component matrix for the objects is constrained to have zero entries for all the components that are not associated to the cluster to which the object in question belongs.

2.3. Relations with other models

In this section, we describe the differences and similarities between Clusterwise Parafac and other models for three-way data that combine clustering with dimension reduction. Within these models, based on the work of[28], a distinction can be drawn between models that use dimension reduction to shed light on between-cluster differences in mean level (i.e., the components model the differences in the mean profiles of the different clusters) and models that use components to gain insight into between-cluster differences in covariation (i.e., the covariation within the clusters is described through cluster-specific components).

As the proposed model clusters the elements of one mode and models the data in different clusters with separate Parafac models, our new model clearly belongs to the second type of models. To the best of our knowledge, only two other deterministic models belong to this category. Thefirst one is the ParaFac with Optimally Clustered Variables (PFOCV,[29]) model. In PFOCV, an optimal partitioning of the elements of one mode is identified, together with the Parafac components that optimally correspond to the resulting clusters: Each Parafac component corresponds to one cluster only, implying that the number of PFOCV components equals the number of clusters.4It can be concluded that Clusterwise Parafac is a generalization of PFOCV in that the cluster- specific Parafac models may consist of more than one component (and that the number of components per cluster needs to be determined during the analysis). The second model is Clusterwise SCA of which different variants have been proposed5 (see [30,27,31,32]). Like Clusterwise Parafac, the latter model clusters the elements of one mode. Within each cluster, however, Clusterwise SCA models the data with Simultaneous Component Analysis (SCA; [33–41]), whereas Clusterwise Parafac uses Parafac instead; the former model is a more flexible model than the latter one because within a cluster, Clusterwise Parafac restricts the component scores to be equal across the elements of the clustered mode whereas Clusterwise SCA allows the component scores to differ across all elements of the clustered mode.

Deterministic models that focus on between-cluster differences in mean level were proposed by[42,43]. In these models, which can be conceived as three-way generalizations of Reduced K-means[44], also called Projection Pursuit Clustering [45,46], or of Factorial K-means [47,48], a clustering of the elements of one mode is combined with a dimensional reduction of the elements of the other modes. The main difference between these models and the models that focus on between-cluster differences in covariation is that in the first the clustering determines which elements of the clustered mode have identical scores on all extracted components, whereas in the latter the clustering indicates which subset of the components is relevant for summarizing the data of a particular element of the clustered elements (while allowing the scores on these components to differ across elements of the same cluster).

2Without loss of generality, the elements of the object mode (i.e., objects) will be clustered.

3However, allowing the number of components to vary across clusters implies a non- trivial model extension. In particular, using an ALS algorithm tofit the loss function (Eq.(3)) to the model in Eq.(1)with a varying number of components may not give the required result. In this regard,[27]show that an ALS algorithm for a Clusterwise SCA model with cluster-specific R values has the tendency to erroneously assign most of the data blocks to the cluster(s) with the largest number of components as this yields a larger increase in model fit (i.e., more components yield more modeling freedom). This phenomenon especially occurs when the clusters are relatively difficult to distinguish, which, among others, may be caused by a high amount of noise variance. Moreover, allowing the number of components to vary across clusters also complicates model selection as the optimal number of components needs to be determined for each cluster separately.

4Note that[29]also present the ParaFac with Clustered Variables (PFCV) model in which the clustering under consideration is given beforehand.

5Note that Clusterwise SCA was proposed to model multivariate multiblock data and that three-way data can be conceived as a special case of this type of data[21].

(3)

3. Data analysis 3.1. Aim

The aim of a Clusterwise Parafac analysis with Q clusters and R components of a data array X is to estimate P, Aq, Bq, and Dk(q = 1…Q and k = 1…K) such that (1) the value of the loss function

f ¼XK

k¼1

Xk−Mk

k k2F;

f ¼XK

k¼1

‖Xk−XQ

q¼1

pkqAqDkBTq

2F

ð3Þ

is minimized, with ||…||Fdenoting the Frobenius norm (i.e., the square root of the sum of squared values) and (2) the model array M can be perfectly represented by a Clusterwise Parafac model with Q clusters and R components.

3.2. Algorithm

In order tofit a Clusterwise Parafac model with Q clusters and R components to a given I × J × K data array X at hand, the loss function in (Eq. (3)) is minimized. To this end, an alternating least squares (ALS) algorithm will be adopted (for an ALS algorithm for the original Parafac model, see[49]). Because an ALS algorithm may end up in a local rather than the global optimum, it is advised to complement the algorithm with a multi-start procedure, along with a smart way to obtain initial parameter estimates that are already of a“high quality”.

3.2.1. ALS algorithm

Because there exist many different ways of assigning the K objects to Q clusters [i.e., a partition problem is NP-hard, see, e.g., 50,51], determining the global optimum of the Clusterwise Parafac loss function in Eq.(3)is not an easy task, except in those cases where K is small (i.e., Kb20). To this end, we will use an ALS procedure [for more information on ALS, see 52,53] that alternates between updating the cluster memberships of the objects (in partition matrix P) and re-estimating the (cluster-specific) Parafac parameters (i.e., Aq, Bq, and C), until none of these updates result in an improvement in the loss function value.

In particular, the Clusterwise Parafac ALS algorithm consists of the following four steps:

1. Generate an initial partition matrix P by assigning the K objects to one of the Q clusters, without allowing clusters to be empty.

2. Estimate the object component matrix C and the cluster-specific variable and source component matrices Aqand Bq. To this end, perform a separate Parafac on each cluster-specific data matrix Xq

(of size I × J × Kq) that is obtained by only retaining the data slices of X of the objects that belong to cluster q. To estimate these cluster-specific Parafac models, an ALS algorithm is used [49] in which the component matrices Aq, Bq, and the relevant part of C (i.e., only the component scores of the objects that belong to cluster q are updated) are alternatingly re-estimated conditionally upon the other component matrices until no improvement in the loss function is observed [see54–56].6 At the end, compute the loss function value (Eq.(3)).

3. Update the partition matrix P row-wise (i.e., object by object) and re- estimate C, Aq, and Bq(q = 1,…, Q) as in step 2. To determine the optimal cluster membership of object k, for each cluster q, the optimal object component scores for object k given Aqand Bqare

computed by means of linear regression (for more information, see [20]). Next, for each cluster q, the partition criterion Lkq= ||Xk− Aq

DkqBqT||F2is computed, with Dkqbeing a diagonal matrix that contains on its diagonal the component scores for object k when assigned to cluster q. The value for the partition criterion Lkqdenotes the extent to which object k does notfit in cluster q. Finally, object k is re- assigned to the cluster q for which Lkqis minimal (i.e.,^pk= arg minq

Lkq, with^pk denoting the cluster to which object k is assigned to).

After updating the cluster membership of each object, it may occur that one (or more) cluster(s) is empty, which may indicate that there are less clusters in the data than the pre-specified number Q.

In the case where one wants to avoid empty clusters, one may optionally apply the following procedure (a similar procedure is also used to avoid empty clusters in k-means clustering, see[58]):

(a) assign the object thatfits its current cluster the worst to (one of) the empty cluster(s) and re-estimate the component matrices as in step 2, and (b) repeat the procedure in (a) until all clusters are non-empty. When adopting this procedure to avoid empty clusters in cases in which Q is specified too large, the worst fitting object(s) from the Q-1 (Q-2, …) solution will be assigned to a separate cluster.7

4. Compute the loss function value (Eq.(3)). When it has improved (i.e., decreased), return to step 3, otherwise the algorithm has converged.

3.2.2. Multi-start procedure

Because the above described ALS algorithm may end up in a local rather than the global optimum, a multi-start procedure is adopted.

This procedure consists of running the Clusterwise Parafac ALS algorithm with different (pseudo-) random and rational initializations of the partition matrix P[59,60]; the solution with the lowest value on loss function (Eq.(3)) is retained. Because many suboptimal solutions for the Clusterwise Parafac loss function may exist, determining a set of good (“high quality”) initial object partitions is of utmost importance.

In particular, V high quality object partitions can be generated by means of the following procedure:

1. Obtain a rational initial object partition matrix Prational by, first, performing a PCA with R components on each Xk(resulting in a variable and a source component matrix Akand Bkfor each object k). Next, determine the similarity between each pair of obtained variable and source component matrices Akand Bk(k = 1 …K).8 Finally, perform an (unweighted) average linkage hierarchical cluster analysis[62]on the K × K matrix of similarity coefficients, and obtain a (rational) object partition by cutting the resulting tree at Q clusters.9

2. Starting from Prational, generate V × 5 different pseudo-random partition matrices Ppseudo-random (with no empty clusters) by re- assigning each object to another cluster with a probability equal to .15 (with all‘other’ clusters having the same probability of being assigned to). Additionally, generate X random object partitions

6To minimize the risk of retaining a suboptimal solution for the (cluster-specific) Parafac parameters, a multi-start procedure is used in which the alternating Parafac algorithm is repeated many times, each time starting with other rationally or randomly determined parameter values, and the best encountered solution is selected as thefinal one (for more information, see[57]).

7In these cases, the following features of the obtained solution may indicate that Q is specified too large: (1) the obtained solution has a singleton cluster(s), and/or (2) for two (or more) clusters the obtained Aqand Bqcomponent matrices are very similar. Note, however, that a singleton cluster will also occur when the Parafac components underlying a single object are very distinct from the components underlying the other (clusters of) objects.

8In particular, to determine the similarity between object k and object k′, first, Ak(Bk) is rotated obliquely towards Ak′(Bk′) in a least squares sense, resulting in Ak′rotated(Bk′rotated).

Next, for each associated component in Akand Ak′rotated(Bkand Bk′rotated) the Tucker congruence coefficient[61]is calculated and the average of all these coefficients is computed.

9Average linkage hierarchical clustering was adopted because (1) it is a standard method for clustering objects on the basis of similarities, (2) it consistently performs among the best when comparing different hierarchical clustering methods[63–69], (3) solutions for all possible numbers of clusters are obtained at once (i.e., by cutting the tree at different places), and (4) it is computationally fast. Regarding the latter, note that average linkage hierarchical clustering is often used to obtain in a fast way a good initial start for other clustering procedures, like k-means[70,65].

(4)

Prandom(with no empty clusters), with each object having the same probability of being assigned to each cluster.

3. Estimate for Prational, each pseudo-random Ppseudo-random, and each random Prandom, the associated Clusterwise Parafac model parameters (as in step 2, see above) and calculate the resulting loss function value (Eq.(3)) for each starting solution. Next, sort all 1 + (V × 5) + X solutions according to their loss function value and retain the best V solutions.

The Clusterwise Parafac algorithm has been implemented in MATLAB-code (version 8) and is available upon request from thefirst author. For estimating the cluster-specific Parafac models, code from the N-way Toolbox in MATLAB[57]has been used.

3.3. Model selection

Usually the optimal number of clusters Q and the optimal number of components R of a Clusterwise Parafac model underlying a data array at hand are unknown in advance. As a way out, different Clusterwise Parafac analyses with increasing numbers of clusters q (i.e., q = 1,…, Qmax) and components r (i.e., r = 1,…, Rmax) are conducted. An optimal model is selected by looking for the model with the best balance between, on the one hand,fit to the data (i.e., the loss function value), and, on the other hand, the complexity of the model (i.e., the number of clusters and components). To this end, different model selection strategies may be used, like CORCONDIA[71]and CHull, a numerical convex hull based method[72–74].

Comparing different model selection strategies (a.o., CHull),[27]

shows that a specific sequential method clearly outperforms the others.

This sequential method, which may be conceived as a generalization of the well-known scree test[75], consists of the following two-step procedure[31]: (1) determine the optimal number of clusters Q, and (2) given the selected optimal number of clusters Q, identify the optimal number of components R (for a similar model selection approach in the context of Clusterwise HICLAS, see[76]). In particular, thefirst step of this procedure consists of computing for each value q (q = 2,…, Qmax− 1), given different numbers of components r (r = 1,…, Rmax), the following scree ratio srq|r:

srqjr¼SSQq;r−SSQq−1;r

SSQqþ1;r−SSQq;r; ð4Þ

with SSQq,rdenoting the loss function value (Eq.(3)) for a Clusterwise Parafac model with q clusters and r components. Note that for the smallest (i.e., q = 1) and the largest (i.e., q = Qmax) number of clusters q, the scree ratio srq|ris not defined (implying that the procedure cannot select these numbers of clusters). Next, for each value q (q = 2,…, Qmax− 1), the mean scree ratio srqis calculated by averaging srq|rover the different number of components (i.e., srq¼∑Rr¼1maxsrqjr

Rmax ), and the optimal number of clusters Q is determined as the q for which srqis maximal.

In the second step of the procedure, the optimal number of components R is determined by calculating for different values of r (r = 2,…, Rmax− 1) the following scree ratio srr|Q, with Q being the optimal number of clusters as selected in thefirst step:

srrjQ¼SSQr;Q−SSQr−1;Q

SSQrþ1;Q−SSQr;Q: ð5Þ

Note that the smallest (i.e., r=1) and largest (i.e., r=Rmax) value of r cannot be selected. Next, the optimal number of components R is identified as the value r for which srr|Qis maximal. It should be noted that thefinal decision regarding the model to be retained should also be based on the interpretability and stability of the Clusterwise Parafac solution.

4. Simulation study 4.1. Problem

In this section, we will present a simulation study to evaluate the performance of the Clusterwise Parafac algorithm. At this point, we are interested in five performance aspects: optimization, recovery, computation time, number of iterations performed, and degeneracy of the obtained solutions. With regard to optimization, we will examine how sensitive the algorithm is to local minima. Concerning recovery, we will determine the degree to which the algorithm succeeds in disclosing the true object partition and cluster-specific component structure underlying the data. Moreover, we will investigate whether and how both aspects depend onfive data characteristics. The first three characteristics that will be investigated pertain to the partition of the objects: (1) the number of underlying clusters, (2) the cluster sizes, and (3) the degree of congruence (i.e., similarity) between the cluster- specific variable and source components. Regarding these three characteristics, we expect that algorithmic performance will deteriorate:

(1) when the number of underlying clusters increases [77,78,30,27], (2) when the clusters are of different sizes [79,78,60,27], and/or (3) when there is a large amount of congruence between the cluster- specific component matrices [80,32]. The fourth manipulated factor is the number of underlying components in the cluster-specific Parafac models. Regarding this factor, which reflects the complexity of the underlying Parafac model(s), we expect that the algorithmic performance will decrease with an increasing number of components [81–83,30,27]. Finally, thefifth factor pertains to the amount of noise in the data, for which we hypothesize that algorithmic performance will decrease when the data become more noisy[77,83,30,27].

4.2. Design and procedure

In order to not have an overly complex design, in the simulation study, the size of the data set was kept constant. In particular, the number of variables wasfixed at 15 (I = 15), the number of sources at 30 (J = 30), and the number of objects at 120 (K = 120). Further, the five factors that were introduced above were systematically manipulated in a completely randomized five-factorial design in which all factors were considered random:

1. the number of clusters, Q, at 2 levels: 2 and 4;

2. the cluster sizes, at 3 levels [see78]: equal (objects are equally distributed across clusters); unequal with minority (one small cluster having 10% of the objects and the remaining objects being equally distributed over the other clusters); unequal with majority (one large cluster containing 70% of the objects and the remaining objects being equally distributed over the other clusters);

3. the degree of congruence between the cluster-specific component matrices Aqand Bq(q = 1,…, Q). The congruence for Aqand Bqwas manipulated independently at 2 levels (i.e., low congruence and moderate congruence), resulting in the following 4 levels for the congruence factor: (1) low congruence for Aqand low congruence for Bq, (2) low congruence for Aqand moderate congruence for Bq, (3) moderate congruence for Aqand low congruence for Bq, and (4) moderate congruence for Aqand moderate congruence for Bq; 4. the number of components of the cluster-specific Parafac models, R,

at 2 levels: 2 and 4;

5. the amount of noise in the data,ε, at 5 levels: .10, .30, .50, .70, and .90.

For each combination of the levels of thefive manipulated factors (i.e., cell of the design), 10 data sets X were constructed by means of the following procedure: First, a true partition matrix Ptrue was constructed by computing the number of objects that belong to each cluster (given the number of clusters and the cluster sizes) and assigning at random the correct number of objects to each cluster.

Next, a true object component matrix Ctrue, and true cluster-specific

(5)

variable Aqtrueand source Bqtruecomponent matrices (q = 1… Q) were generated. The true object component matrix Ctruewas obtained by independently drawing entries from a uniform distribution from the interval [−1 1]. To manipulate the amount of congruence between the different Aqtrueand Bqtrue(q = 1… Q), first, a (common) base variable Abaseand source Bbasecomponent matrix was generated by drawing entries from U (−1,1). Next, matrices Aqtempand Bqtempwere simulated by drawing entries from U (− 2.50, 2.50) or U (−.60, .60)10, and Aqtemp

and Bqtemp

were added to Abaseand Bbase, respectively (e.g., Aqtrue

= Abase+ Aqtemp). As a consequence, a set of Aqtrueand Bqbasematrices (q = 1

… Q) that are moderately and lowly congruent were obtained.11Next, true matrices Tk(k = 1… K) were computed by combining Ptrue, Aqtrue

, Bqtrue, and Ctrueby the Clusterwise Parafac decomposition rule inEq. (2).

Finally, for each true matrix Tk, a data matrix Xkwas constructed by adding a noise matrix Ek(of size I × J) to Tk. To manipulate the amount of noise in the data, each noise matrix Ekwas constructed by,first, generating a noise matrix Ekstart(of size I × J) by independently drawing numbers from N(0,1), and, next, rescaling Ekstartto ensure∑k∈Cq kð Þk kEk 2F

k∈Cq kð Þk kXk 2F

being equal toε for each cluster q, with Cq(k)denoting the set of objects that belong to the same cluster to which object k belongs.

Fully crossing all design factors, 10 (replication) × 2 (number of clusters) × 3 (cluster sizes) × 4 (congruence) × 2 (number of components) × 5 (amount of noise) = 2400 different data arrays X were obtained. Subsequently, a Clusterwise Parafac analysis with the true number of clusters Q (not allowing clusters to be empty) and the true number of components R was performed on each data array X with 25 starts. The starts were obtained by selecting the best 25 initial partition matrices P among (1) a rationally determined P, (2) 125 pseudo-random P (with the probability of re-allocating an object to another cluster being equal to .15), and (3) 25 random P (see Section 3.2). The simulation study was run on a supercomputer consisting of INTEL XEON L5420 processors with a clock frequency of 2.5 GHz and with 8 GB RAM.

4.3. Results

4.3.1. Optimization performance: sensitivity to local minima

The aim of this section is to investigate whether the Clusterwise Parafac algorithm is capable of identifying the global optimum of the loss function (Eq. (3)). From the moment, however, that noise is added to the data, the global optimum of the loss function is unknown.

A way out consists of defining a proxy for the global optimum, which can be conceived as our best guess of the loss function value of the global optimal solution. We define the proxy as the lowest loss function value encountered among the following two Clusterwise Parafac solutions: (1) the true solution underlying the data (i.e., Ptrue, Aqtrue, Bqtrue, and Ctrue), because this true solution is always a valid solution with Q clusters and R components for the data at hand, and (2) the solution

obtained by running the Clusterwise Parafac algorithm using the true clustering Ptrueas start. Note that in most cases the second option will lead to a tighter proxy (i.e., a lower loss function value) than thefirst one. We consider a Clusterwise Parafac solution as suboptimal when its loss value exceeds the proxy. This appeared to be the case for 578 out of the 2400 data sets (i.e., 24.08%). InTable 1, which presents the percentage of data sets for which a suboptimal solution is encountered for all levels of thefive manipulated factors, one can see that local minima mostly occur in the conditions with four components, when there is a large amount of noise, and when there is a large cluster together with one (or more) small cluster(s).

4.3.2. Recovery performance

In this subsection, we will evaluate the recovery performance of the Clusterwise Parafac algorithm regarding (1) the object partition, and (2) the cluster-specific variable and source component matrices.

4.3.2.1. Recovery of the true partition of the objects. To determine the degree to which the underlying object partition has been recovered by the Clusterwise Parafac algorithm, we calculated the Adjusted Rand Index [ARI;84] between the true object partition (i.e., Ptrue) and the object partition retained by the algorithm (i.e., P). An ARI value of 1 is encountered when both partitions are identical, whereas ARI equals 0 when recovery is at chance level.

Across all 2400 data sets, the mean ARI equals .9315 with a standard deviation of .1553. Moreover, a perfect recovery of the true object partition (i.e., ARI = 1) was encountered for 1545 out of the 2400 data sets (64.38%). This implies that the Clusterwise Parafac algorithm recovers the underlying object partition to a very large extent. To investigate whether the recovery of the true partition depends on the manipulated data characteristics, for all levels of thefive manipulated factors, the mean ARI value is presented inTable 2. In this table, one can see that the cluster recovery decreases when the amount of noise increases, the number of clusters increases, the degree of congruence among cluster-specific component matrices increases, and when there is one small cluster with one (or more) large(r) cluster(s). Because there is only a limited variation in ARI values, no analysis of variance has been performed.

4.3.2.2. Recovery of the true cluster-specific variable and source component matrices. The extent to which the Clusterwise Parafac algorithm recovers the true cluster-specific variable and source component matrices Aqand Bqwas determined by,first, calculating, for each cluster q, the Tucker congruence coefficient[61]θqAqB) between the true Aqtrue(Bqtrue) and the estimated Aq(Bq). To account for the within-cluster permutation freedom of the Parafac components, for each cluster q, that permutation

10In a pilot study it was found that for the data characteristics used in the simulation study adopting a value of 2.50 for c yields a mean Tucker congruence[61]value around .15 (i.e., a low congruence), whereas a c-value of .60 results in a mean Tucker congruence value around .75 (i.e., a moderate congruence).

11To determine the amount of congruence between the resulting cluster-specific component matrices Aqtrue

and Bqtrue

at the four levels of the congruence factor, the Tucker congruence coefficient (with 1 indicating perfect congruence/similarity; see [61]) between the columns of each couple of Aqtrue(Bqtrue) matrices was computed, after taking the permutation freedom of the Parafac components into account. When averaging this Tucker congruence coefficient over all components and all couples of cluster-specific variable (source) component matrices, the following mean congruence scores are obtained for the four levels of the congruence factor (for Aqand Bq, respectively):

(1) .1879 (SD = .1072) and .1401 (SD = .0793) for the low Aqcongruence–low Bq

congruence condition, (2) .1265 (SD = .1201) and .7328 (SD = .0369) for the low Aq

congruence–moderate Bqcongruence condition, (3) .7331 (SD = .0561) and .1379 (SD = .0887) for the moderate Aqcongruence–low Bqcongruence condition, and (4) .7287 (SD = .0535) and .7347 (SD = .0374) for the moderate Aqcongruence–moderate Bqcongruence condition.

Table 1

Percentage of data sets for which a suboptimal solution was encountered (i.e., proxy miss) for all levels of the manipulated factors.

Factor Levels Percentage

proxy miss

Number of clusters 2 22.75

4 25.42

Cluster sizes Equal sized clusters 24.00

One small cluster 17.13

One large cluster 31.13

Number of components 2 14.00

4 34.17

Amount of noise in the data .10 4.79

.30 10.62

.50 17.08

.70 46.46

.90 41.46

Degree of congruence between the cluster-specific component matrices

Low Bq–low Cq 24.67

Low Bq–moderate Cq 23.67 Moderate Bq–low Cq 22.67 Moderate Bq–moderate Cq 25.33

(6)

of the components was selected that optimizesθqAqB. Next, an overall θallAallB)-statistic was obtained by averaging the cluster- specific θqAqB) across all clusters:

θAall¼ XQ

q¼1θ Atrueq ; Aq

 

Q ¼

XQ q¼1θqA

Q ;

θBall¼ XQ

q¼1θ Btrueq ; Bq

 

Q ¼

XQ

q¼1θBq

Q :

ð6Þ

The between-cluster permutation freedom was taken into account by trying all cluster permutations and selecting the cluster permutation that yields the largest value forθABall¼θAallþθ2Ball. AθallABvalue of 1 indicates that the recovery of the true cluster-specific variable and source components is perfect, whereas a value of 0 implies that there is no recovery at all.

Across all data sets, the meanθallABequals .9712 with a standard deviation of .0663, indicating that the Clusterwise Parafac algorithm recovers the cluster-specific variable and source components to a very large extent. In order to study how the recovery of Aqtrueand Bqtruevaries as a function of the manipulated data characteristics, the meanθallAB

value for each level of thefive manipulated factors is displayed in Table 2. From this table, although the differences are small, it appears that the recovery of the cluster-specific variable and source component matrices decreases when the optimization problem becomes more complex (i.e., a larger number of clusters and a larger amount of noise in the data) and when there is one small and one (or more) large(r) cluster(s) in the data. Because there was too little variation among the θallABvalues (i.e., more than 86% of the values is larger than .95), no analysis of variance has been performed.

4.3.3. Computation time, number of performed iterations, and the occurrence of degenerate Parafac solutions

In the simulation study, the mean computation time for one simulated data set equals 3.23 h with a standard deviation of 2.41 h.

To investigate how the computation time depends on the manipulated data characteristics, an analysis of variance was performed with the computation time as the dependent variable and the manipulated factors as independent variables. Only taking effects into account with an intraclass correlation ^pI [85,86]larger than .10, it appears that the computation time increases with increasing numbers of clusters (^pI= .28), amounts of noise in the data (^pI= .20), and numbers of components (^pI = .14). In particular, taking four instead of two clusters implies the computation time to double, whereas going from

two to four components lengthens the computation time with 70%.

For the amount of noiseε equaling .10, .30, .50, .70, and .90, the mean computation time equals 2.22, 2.23, 2.29, 3.60, and 5.81 h, respectively.

These main effects, however, are qualified by a considerable interaction between the number of clusters and the amount of noise (^pI= .11). In particular, as one can see inTable 3, the increase in computation time for higher number of clusters is more pronounced when the data contain larger amounts of noise.

The number of iterations that is performed by the algorithm is mainly influenced by the amount of noise (^pI= .22) and the number of clusters (^pI= .10). In particular, the algorithm iterates more when the amount of noise (i.e., on average, 2.89, 3.30, 3.86, 4.56, and 6.56 iterations forε=.10, .30, .50, .70, and .90, respectively) and the number of clusters (i.e., on average, 3.54 and 4.93 iterations for 2 and 4 clusters, respectively) increase.

To determine whether some of the obtained cluster-specific Parafac models are degenerate (i.e., some components are proportional to each other), we based ourselves on the degeneracy criteria discussed by [87,88]: (1) the triple cosine product [89,29,90–92]is smaller than−.85, (2) the smallest eigenvalue of the triple-cosine matrix (i.e., the R × R matrix with the triple cosine products between the R components as entries) is smaller than .50, and (3) the condition number (i.e., the ratio of the largest eigenvalue to the smallest eigenvalue) of the triple-cosine matrix is larger than 5. More information on the degeneracy of a Parafac solution and ways to deal with this problem can be found in [2,93,24,94–97,49]. Applying these criteria in our simulation study reveals that almost all our obtained solutions are non- degenerate. In particular, only for 14 data sets (.58%) the minimal triple cosine product across the clusters is smaller than −.85.

Further, only for 35 data sets (1.46%) is the smallest eigenvalue smaller than .50, whereas a condition number larger than 5 is only encountered for 31 data sets (1.29%). Note that these few degenerate solutions all belong to the simulation conditions with four clusters.

4.3.4. Discussion of the results

From the simulation results presented above, it appears that, in general, the Clusterwise Parafac algorithm performs well with respect to optimizing the loss function when using 25“high quality” multi- starts (see Section 3.2), although local minima problems may be encountered when the data contain an excessive amount of noise (i.e., more than 70%). Therefore, when the optimization problem becomes more difficult (i.e., larger number of clusters/components and larger amounts of noise), we advise to increase the number of starts that is used throughout the algorithm. Moreover, the algorithm recovers both the true object partition and the true cluster-specific variable and source components to a very large extent. As such, it can be concluded that, when the optimal number of clusters and components is known, the Clusterwise Parafac algorithm with 25 starts delivers solutions of a good quality. One may wonder what would happen if we carry out the analysis with too few or too many clusters. In this regard,[32]show for Clusterwise SCA that when too few clusters are estimated, two true clusters get cleanly fused into a single cluster, whereas extracting too Table 2

Mean ARI value (and standard deviation) and meanθallABvalue for all levels of the manipulated factors.

Factor Levels Mean ARI

(standard deviation of ARI)

Mean θallAB

Number of clusters 2 .9638 (.1107) .9893

4 .8991 (.1841) .9532

Cluster sizes Equal sized clusters .9683 (.0829) .9907 One small cluster .8855 (.1968) .9465 One large cluster .9406 (.1524) .9764

Number of components 2 .9221 (.1592) .9770

4 .9408 (.1507) .9655

Amount of noise in the data

.10 .9836 (.0829) .9929

.30 .9632 (.1276) .9852

.50 .9592 (.1340) .9824

.70 .9444 (.1293) .9770

.90 .8070 (.2063) .9186

Congruence between the cluster-specific component matrices

Low Bq–low Cq .9366 (.1482) .9651 Low Bq–moderate Cq .9383 (.1419) .9733 Moderate Bq–low Cq .9379 (.1546) .9720 Moderate Bq–moderate Cq .9131 (.1732) .9744

Table 3

Mean computation time (and standard deviation) in hours for Clusterwise Parafac analyses (with 25 multi-starts, which are the best ones out of 151 initial partitionings) for different amounts of noise in the data and varying numbers of clusters.

Amount of noiseε 2 clusters 4 clusters Overall

.10 1.55 (.41) 2.88 (.89) 2.22 (.96)

.30 1.56 (.39) 2.90 (.91) 2.23 (.97)

.50 1.57 (.39) 3.01 (.96) 2.29 (1.03)

.70 2.24 (.65) 4.97 (1.77) 3.60 (1.91)

.90 3.51 (1.29) 8.11 (3.65) 5.81 (3.58)

overall 2.09 (1.05) 4.37 (2.81) 3.23 (2.41)

(7)

many clusters results in one true cluster being split into two smaller clusters. Given the close relationship between both methods (see Section 2.3and[21]), a similar result may be expected for Clusterwise Parafac.

5. Illustrative application

In this section, Clusterwise Parafac will be illustrated by re-analyzing sensory profiling data. The goal of analyzing such data is disclosing the experience dimensions that underlie the attribute ratings of a set of food samples by different panelists (i.e., QDA; Quantitative Descriptive Analysis with fixed vocabulary profiling). Herewith, the similarities and differences among the food samples can be uncovered by studying the sample scores on the dimensions[4]. However, as argued in the Introduction, it is very likely that the ratings of different groups of food samples may rely on different experience dimensions. Such qualitative differences between sample groups can be revealed by applying Clusterwise Parafac while clustering the sample mode.

The data set that we will analyze was collected by[98]and pertains to 10 different cheese types (with one cheese type being used twice– as an internal reference– to test the stability of the evaluations of the panelists across the different sessions). For each of the 10 cheese types, 3 samples were created, resulting in 30 cheese samples. Each of the 30 cheese samples was rated by 8 trained panelists on 23 attributes, which refer to textural, appearance, and taste related features that can be observed by smelling, by sight, by tasting, and by touching (an overview of the attributes can be found inTable 7). To score the attributes, a horizontal 15cm unstructured line scale was used, resulting in scores between 0 and 15. More information regarding the data collection can be found in[98] and [4]; the data were retrieved from http://www.models.kvl.dk/Cream.

The data were pre-processed by,first, centering the data across the cheese samples (i.e., for each combination of an attribute and a panelist the mean across the cheese samples equals zero) and, next, scaling the data per attribute (i.e., across all combinations of a cheese sample and a panelist) to a sum of squares of 1 (for more information regarding pre-processing of three-way data, see [6,99,93]). Next, different Clusterwise Parafac analyses (with 25 starts, being the best ones out of 1 rational start, 125 pseudo-random starts, and 25 random starts) were performed on the pre-processed data, with the number of clusters ranging from one to six (no empty clusters allowed) and the number of components from one to four. InTable 4, one can see that a Clusterwise Parafac analysis gets computationally cumbersome when the number of clusters and especially the number of components becomes large. Some of the latter solutions (i.e., many clusters and/or components) suffer from degeneracy problems.

In order to select an optimal model, we applied the generalized scree test[31], as explained inSection 3.3, to the obtained loss function values (seeFig. 1). InTable 5, in which the srq|rvalues (i.e., second tofifth column) are displayed, together with the srqvalues (i.e.,final column), one can see that the optimal number of clusters equals three as the srq

value is the largest for Q = 3. When only considering the solutions

with Q = 3 clusters, the scree ratio values srq|Q = 3equal 1.3800 and 1.0237 for the solution with 2 and 3 components, respectively. Hence, we selected the solution with three clusters and two components.

Each cluster-specific Parafac model of this solution is non-degenerate (i.e., across all clusters, the smallest triple cosine product equals

−.0806, the smallest eigenvalue is .9194, and the largest condition number equals 1.18).

In Table 6, in which the obtained cheese sample partition is presented, one can see that thefirst cluster consists of (all the samples of) the Standard full fat cream cheese (34%) and the Commercial cream cheese D (D-CHO). The second cluster subsumes the Medium fat reduced (24%) and the Maximum fat reduced cream cheese (16%), together with the Commercial cream cheese B (B-Prot), the Prototype cream cheese (P), and the Prototype cream cheese with added Butter Aroma (P + Aroma). The third cluster,finally, collects both Commercial cream cheeses A (A-Prot) and the Commercial cream cheese C (C-CHO).

The three samples of the same cheese type are always assigned to the same cluster, except for one sample of the Medium fat reduced cream cheese 24% (i.e., assigned to cluster 3 instead of cluster 2) and one sample of the Commercial cream cheese A (i.e., assigned to cluster 2 instead of cluster 3).

When studying the cluster-specific attribute component scores, which are presented inTable 7, it appears that the texture (i.e., Hand- and Mouth-resistance and Mouth-firm versus Mouth-melt down) and the appearance (i.e., Eyes-white versus Eyes-yellow) of the cheeses are the most important dimensions to discriminate among the 33%

and the D-CHO samples (i.e.,first cluster). In particular, inTable 6and Fig. 2, in which the sample component scores are displayed (for each cluster), one can see that the 33% cheese is considered soft (i.e., Mouth- melt down) and has a yellow appearance (i.e., large negative scores on both components), whereas the D-CHO cheese is white and has afirm texture (i.e., large positive component scores). For the second cluster (i.e., 16%, 24%, B-Prot, P, P + Aroma), the texture of the products is again

Table 4

Computation time in hours for different Clusterwise Parafac analyses of the cheese data (with 25 starts, which are the best ones out of 151 initial partitionings).

Number of clusters 1 component 2 components 3 components 4 components

1 .00 .01 .02 .03

2 .48 2.47 4.68 5.43

3 .85 3.09 5.38 5.75

4 1.11 3.55 5.53 9.41

5 1.26 3.75 7.34 9.49

6 1.66 3.82 7.36 8.23

1 2 3 4

12 14 16 18

Number of components

Sum of squared residuals

Scree plot

1 cluster 2 clusters 3 clusters 4 clusters 5 clusters 6 clusters

Fig. 1. Loss function values for thefitted Clusterwise Parafac solutions with the number of clusters varying from one to six and the number of components ranging from one to four for the cheese data.

Table 5

Scree ratios srq|rfor the number of clusters q (q = 2,…, Qmax− 1) given the number of components r (r = 1,…, Rmax) and average scree ratio srq over the numbers of components for the cheese data. The largest scree ratio in each column is highlighted in boldface.

Number of clusters (Q)

1 component

2 components

3 components

4 components

Mean over components (srq)

2 1.60 1.52 1.92 1.56 1.32

3 2.71 1.53 .95 1.66 1.37

4 1.08 .79 1.78 .89 .91

5 1.47 1.36 1.27 1.25 1.07

Referenties

GERELATEERDE DOCUMENTEN

These model parameters were then used for the prediction of the 2007 antibody titers (the inde- pendent test set): Component scores were derived from the 2007 data using the

Specifically, in OC-SCA-IND, the ‘status’ i.e., common versus some degree of cluster-specificity of the different components becomes clear when looking at the clustering matrix,

Specifically, 2M-KSC assigns the persons to a few person clusters and the symp- toms to a few symptom clusters and imposes that the time profiles that correspond to a specific

As to the performance of the CHull model selection procedure, we conclude that CHull can be a useful aid to indicate the number of zero vectors p, but that there is room for

In the PCovR case, the four model selection strategies that were introduced in Section 4.3 (ML_SCR, SCR_ACV, ML_RCV and CV) were applied. When α is tuned through cross-validation,

This application demonstrated that using clusterwise SCA to investigate what is causing a lack of measurement invariance makes sense, because (1) the fit of the multigroup CFA mod-

We see two possible pitfalls, however: (a) estimating all components as cluster-specific can affect the recovery of the clustering in case the data contain a lot of error and the

In Tucker3, the situation is somewhat more complex, as individual differences in the used dimensions may be represented by the component scores of the panelists as well as by the