- Methodology article
- Open Access
- Published:

# A simple procedure for the comparison of covariance matrices

*BMC Evolutionary Biology***volume 12**, Article number: 222 (2012)

## Abstract

### Background

Comparing the covariation patterns of populations or species is a basic step in the evolutionary analysis of quantitative traits. Here I propose a new, simple method to make this comparison in two population samples that is based on comparing the variance explained in each sample by the eigenvectors of its own covariance matrix with that explained by the covariance matrix eigenvectors of the other sample. The rationale of this procedure is that the matrix eigenvectors of two similar samples would explain similar amounts of variance in the two samples. I use computer simulation and morphological covariance matrices from the two morphs in a marine snail hybrid zone to show how the proposed procedure can be used to measure the contribution of the matrices orientation and shape to the overall differentiation.

### Results

I show how this procedure can detect even modest differences between matrices calculated with moderately sized samples, and how it can be used as the basis for more detailed analyses of the nature of these differences.

### Conclusions

The new procedure constitutes a useful resource for the comparison of covariance matrices. It could fill the gap between procedures resulting in a single, overall measure of differentiation, and analytical methods based on multiple model comparison not providing such a measure.

## Background

Covariance matrices are key tools in the study of the genetics and evolution of quantitative traits. The **G** matrix, containing the additive genetic variances and covariances for a set of characters, summarizes the genetic architecture of traits and determines their short-term response to multivariate selection along with the constraints this response will face. The more easily estimated matrix of phenotypic variances and covariances **P** can be used as a surrogate for **G**, especially in the case of high heritability morphological characters [1–4]. Comparisons between covariance matrices are carried out in the study of a wide array of evolutionary problems, such as the stability of **G** in the presence of selection and drift [5–7], the role of genetic constraints on determining evolutionary trajectories in adaptive radiations [8], the response of genetic architecture to environmental heterogeneity [9], the evolution of phenotypic integration [4, 10], multi-character phenotypic plasticity [11] and sexual dimorphism [12, 13].

Several methods for the comparison of covariance matrices are available (reviewed in [14]). They range from the most mathematically sophisticated, such as those using maximum likelihood [15] or Bayesian frameworks [16], to simple methods that are useful for exploratory analyses and are not dependent on distributional assumptions. The simplest methods [17–20] are based on pair wise comparisons of the matrices’ elements, so that they typically ignore the lack of independence between these values, cannot detect proportionality between matrices, and consider two matrices only. More recent procedures, also using the matrix elements [21], take into account these elements’ lack of independence and permit the joint consideration of several matrices, making it possible to study the contribution of identified external factors to the magnitude of the differentiation.

Other simple procedures [22, 23] are based on comparisons between vectors resulting from the product of the studied matrices and sets of test vectors, their rationale being that similar matrices would produce similar results when multiplied by the same sets of vectors. However, most of these procedures result in a single measure of the divergence between matrices that does not provide information about the nature of this divergence. Such information is provided by CPCA (Common Principal Components Analysis [24]), which uses the Flury [25] hierarchy, a maximum likelihood-based procedure to compare the structure of two or more matrices and sequentially test if the matrices are “unrelated” (have no eigenvectors in common); if they have some number of eigenvectors in common, if they are proportional (have all their eigenvectors in common and their eigenvalues keep the same proportions), and finally if they are equal (have all eigenvectors and eigenvalues in common). Then it determines which of these descriptions best fits to the relationship between the matrices’ structures.

Among the limitations of CPCA are, first, that it is based on the assumption of multivariate normality, and second, that it results in categorical, not continuously varying measures of matrix similarity [26]. The CPCA consists in a series of yes/no comparisons between ordered eigenvectors, which allow testing a full series of hypotheses about the relationship between matrices in a hierarchical way, but idoes not have an associated parameter measuring the degree of similarity, relying only on the results of the significance tests. This limitation can be serious in some situations. Two matrices are declared as “unrelated” when that is the best fit of all null hypotheses tested, but this result does not preclude the existence of any similarities between them [14]. In fact, the procedure may declare two matrices as “unrelated” when many data are available and there is great power to detect differences, even if these differences are trivial from a biological point of view [26, 27]. Thus, there is no simple relationship between matrix similarities measured by CPCA and other matrix comparison procedures [27].

In the present work I propose a new, simple and distribution-free procedure for the exploration of differences between covariance matrices that, in addition to providing a single and continuously varying measure of matrix differentiation, makes it possible to analyse this measure in terms of the contributions of differences in matrix orientation and shape. I use both computer simulation and **P** matrices corresponding to snail morphological measures to compare this procedure with some widely used alternatives. I show that the new procedure has power similar or better than that of the simpler methods, and how it can be used as the basis for more detailed analyses of the nature of the found differences.

### Pairwise matrix comparison

The rationale for the comparison procedure is that, when the covariance matrices of two data samples are similar, the eigenvectors obtained in a principal component analysis of any of them will explain similar amounts of variation in both samples. The degree of similarity can be measured by calculating in each sample the individuals’ values and variances for the eigenvectors obtained in the other sample. Given that **D**_{
1
} and **D**_{
2
} are the matrices with the characters’ measures in the two samples and **X**_{
1
} and **X**_{
2
} the matrices containing in their columns the eigenvectors of these samples’ covariance matrices, the variances of the columns of the products **D**_{
1
}**X**_{
1
} and **D**_{
2
}**X**_{
2
} are the corresponding eigenvalues, i.e., the amounts of variance explained by the original eigenvectors, and those of **D**_{
1
}**X**_{
2
} and **D**_{
2
}**X**_{
1
}, the amounts of variance explained by the eigenvectors from the compared sample. Thus, for each of the *n* (number of variables measured) pairs of eigenvectors obtained in the analysis of the two samples it is possible to calculate v_{i11}, v_{i12}, v_{i21}, and v_{i22}, where v_{i11} is the amount of sample 1 total variance explained by eigenvector i from sample 1, v_{i12} the amount of sample 1 total variance explained when applying eigenvector i from sample 2, and so on. These n sets of four values are the basic items to measure the similarity in covariance between samples. I define three sums:

where S1 is a general measure of differentiation depending on the ability of the eigenvectors from each sample to explain the variation in the other sample; S2 is a measure of the contribution of between-matrix differences in orientation (i.e., differences in orientation between eigenvectors in the same ordinal position in the two matrices) to S1, and S3 that of differences in shape (i.e., differences in the proportion of total variance explained by eigenvectors in the same ordinal position in the two matrices). It can be shown (Appendix 1) that S1 = S2 + S3. Figure 1 shows a graphical interpretation of these statistics. In this figure, row A shows that S3 is the only non-zero component of S1 when the matrices differ in shape but not in orientation, and row B that S2 is the only non-zero component when the matrices differ in orientation but not in shape. Note however that S2 and S3 do not measure changes in orientation or shape as such. What these statistics measure are the consequences of changes in orientation and shape on the more functionally relevant proportions of variance explained, i.e.,. on the amount of variation available in each multidimensional direction. The effects of these changes are not independent. The comparison between Figure 1 rows B and C shows that the impact of changes in orientation on the proportion of variance explained by the eigenvectors of the reciprocal matrix depends on the matrices’ shape. The effect is larger for more “elongate” matrices. Similarly, the comparison between rows A, D and E shows that the effect of changes in shape depends on the matrices’ orientation, and that changes in matrix shape become irrelevant (in this two-variables example) when the first eigenvectors of the two matrices are perpendicular (row D). Finally, row F shows that there are no differences in orientation when one of the matrices is perfectly “round” (i.e., all its eigenvectors explain the same amount of total variance). Despite this complexity, and as shown in the practical application made below, statistics S2 and S3 may be very useful to analyse the nature of the differentiation between two covariance matrices.

The S statistics are easier to compare between studies if they are made to vary between zero and one by making them relative to their maximum possible value. This maximum would occur in the extreme situation in which single eigenvectors explain all variation in each of the compared samples, and the eigenvectors of the two samples are orthogonal. In that case, S1 is equal to 2 times the sum of twice the square of the total variance of the first sample and twice the square of the total variance of the second sample. When the variances explained by each eigenvector are expressed as proportions of the total variance, so that the sum of all the proportions is equal to one, the maximum possible value for S1 is equal to eight. In the computer simulation and real data example shown in this article, the explained variances are expressed as proportions, and the S1, S2 and S3 statistics divided by eight so that they could vary between zero and one.

Figure 2 shows the responses of the three S statistics in a wide range of two-variable situations. S2 is not sensible to changes affecting only matrix shape (A through E, zero degrees case) and S3 is not sensible to changes affecting only matrix orientation (B and C, all relative orientations). Also, and as explained above, the influence of changes in orientation on the proportion of variance explained by the eigenvectors of the compared matrix is heavily dependent on matrix shape. It is large for elongate matrices and modest for rounded ones (compare the maximum values attained by S1 and S2 in B and C). Plots B to E show that the effect of changes in orientation is stronger in high angle ranges than in low angle ones. Finally, plot A shows that, in the same way as the effect of orientation depends on matrix shape, the effect of shape depends on orientation. When matrices differ both in shape and orientation, there is an intermediate angle that minimizes their differences in total variance explained. In addition, when the matrices’ first eigenvectors become perpendicular to each other (far right of all plots), any differences are detected as changes in orientation. These observations emphasise the point made above that what the S statistics measure are differences in the variance explained by two matrices’ eigenvector sets, and therefore in amount of variation in every direction of the multivariate space in the matrices’ original samples. The statistics also measure the influence of matrix shape and orientation on that differentiation, but not differences in shape and orientation directly.

## Results

I contrasted the results obtained with the proposed procedure with those from other widely used ones, namely CPCA and two simpler procedures providing single measures of matrix differentiation: one, the Random Skewers, based on products with test vectors, and the other, the T method, based on the comparison of matrix elements (see Methods). I followed two approaches. First, I studied the procedures’ power and Type I error through computer simulations that considered covariance matrices differing in shape, orientation or both, in different number of variables and sample size situations. Second, I compared their ability to detect differences between covariance matrices of shell measures from different morphs and populations of the seashore snail *Littorina saxatilis*.

### Computer simulation

Figure 3 shows two-variable examples of the four kinds of datasets used in the simulations. The simulation results (Figure 4) show that S2, the T method and RS tended to produce type I errors with frequencies greater than 5% (Figure 4). The average proportions of false positives (calculated across the first row of graphs of the figure, in which a sample of the reference population is compared with another sample for the same population) for the three statistics were 0.096, 0.077 and 0.108, respectively. As expected, the statistic sensitive to changes in orientation, S2, found similar results for comparisons involving matrices diverging only in shape and for comparisons involving matrices from samples of the same population. This was also the case for S3, the statistic sensitive to changes in shape, when applied to populations differing only in orientation. The RS procedure tended to have at least as much power as the most powerful S statistic for the detection of differences in orientation. In fact, it could have been more powerful than the S statistics in the four cases involving changes in orientation, samples of size 100 and more than two variables, because it was able to detect differences in all replicates of these comparisons. In any case, RS seemed to be relatively better for small samples (compare results for sample size 25 with those for sample size 50). However, it was worse than S1 and S3 for the detection of changes in shape. The T method tended to be the least powerful in all situations.

Figure 4 shows also the results obtained with CPCA. These results cannot be directly compared with those of the remaining methods because they were based on a single replicate instead of an average of replicates for each case. However, the results suggest that CPCA could produce an excess of false positives as the number of variables increases (it declared as not equal the matrices from the same population in the seven variables cases). In these simulations, CPCA should declare the matrices as “unrelated” when they differed in orientation, because in that case they would not share any eigenvector in common and as “CPC”, i.e., sharing all eigenvectors but being not proportional when the matrices differed in shape. The procedure tended to correctly detect changes in shape even for small sample sizes, except for the two variables case. Its performance tended to be lower in the detection of changes in orientation in the smaller sized samples. Overall, CPCA tended to agree with the S and RS procedures for the largest (100) sample size.

*Littorina*data

The eigenvectors and eigenvalues of the six samples’ covariance matrices are shown in Table 1. The first eigenvectors always had coefficients of the same sign, as expected for size descriptors [28], and were similar for all matrices. The remaining eigenvectors included coefficients of contrasting signs, as expected if they were measuring variation in shell shape. The bootstrap distributions of the RS, T%, S1, S2 and S3 statistics for 21 comparisons are shown in Figure 5. Not unexpectedly, some basic common trends were observed in all these statistics. Matrices in within-sample comparisons tended to be found more similar than those in within morph comparisons, and these, more than matrices in between morph comparisons. Also, particular between-samples comparisons resulting in large differences when applying a procedure tended to do the same in the remaining procedures. The T method was, of those shown in Figure 5, the least sensitive to matrix differences. In the case of the RS, S1 and S2 statistics, the bootstrap distributions of the within sample and between morphs comparisons had little overlap, showing that these statistics can easily detect rather modest differences (all observed values for the S-statistics measures of differentiation were lower than 0.02 for a range of possible values from 0 to 1, and those for the RS statistic-measures of similarity were higher than 0.9 for a range of possible values from 1 to 0) between matrices using samples of moderate size. The RS method tended to provide a rather definite separation among the results of four kinds of comparisons: within samples, within Rbs, within Sus and between morphs. The profile of the S statistics’ comparisons was rather different from that obtained with the RS procedure, indicating that the two methods are not equivalent and that they consider matrices’ properties in different ways.

At least in the particular example analyzed here, differences related with matrix shape had the largest weight in the overall measure of differentiation S1, as the comparison results profiles of S1 and S3 were the most similar. The statistic S3 found large differences both between morphs and within the Su morph. The largest differences for S3 corresponded always to comparisons involving the matrix of the Sus from location 2, i.e., comparisons 10, 12, 13, 18 and 20 in Figure 4. This suggested that the shape of this matrix had some particularities, and in fact Table 1 shows that the first and second eigenvalues for this sample were unusually small and large, respectively. A more detailed analysis of the S statistics (Figure 6) was consistent with this interpretation: the largest contributions to S3 were made by the first eigenvector and occurred in comparisons involving the Su2 sample.

The statistic S2 found the most striking contrast between kinds of comparisons in Figure 5, differences between morphs being clearly the largest. Thus, a large proportion of the differentiation between the two morphs’ covariance patterns was due to a change in matrix orientation not present in the comparisons within morphs. Figure 5 shows that the largest contributions to S2 were made by the second and third eigenvectors. The examination of these eigenvectors’ coefficients in Table 1 shows that the second eigenvector from the Rbs was similar to the third eigenvector from the Sus, and vice versa. The second eigenvector from the Rbs and the third from the Sus could be roughly described as a contrast of measure 6 against the rest, and the third eigenvector of the Rbs and the second from the Sus as a contrast of measure five against the rest. Thus, the two morphs’ eigenvectors had very similar structure, but the proportion of variance explained by these eigenvectors was clearly different, to the extent that the ranks of the second and third eigenvectors in one morph got reversed in the other morph. This change in matrix orientation introduced large differences in the proportion of variance explained by the reciprocal eigenvectors. This is illustrated in Figure 7, representing the amount of variance explained in a sample by the eigenvectors of the compared sample (v_{i12} and v_{i21} of the S expressions, see above). It can be seen that the differentiation between the Rb and Su matrices is related to a reversal in the variances explained by the second and third eigenvectors. In both morphs, the third eigenvector from the reciprocal morph explains more variation than the second reciprocal eigenvector. The figure shows also that the lowest differentiation was in comparisons involving sample Su3. Again, the inspection of Table 1 results in an easy interpretation: the second and third eigenvalues were very similar in this sample (6.74 and 5.15%), so that reversing the order of the corresponding eigenvectors (as done to a good extent in the comparisons with Rb samples) had very slight effects on the proportion of variance explained. Note in Figure 7 that one of the segments in the circles is nearly horizontal for the comparisons involving sample Su3. Not all results had a straightforward interpretation in terms of eigenvector coefficients. For example, Figure 5 shows that the S-statistics comparisons between the Rb and Su samples from location 3 resulted in markedly lower differentiation than all the remaining between morph comparisons. This shows that the S procedures consider aspects of matrix structure beyond the individual values of eigenvector’s coefficients.

The overall agreement between CPCA and the other procedures found in the simulations was lost in the analysis of *Littorina* data. (upper side of Figure 5). According to the RS, T% and S statistics, the three Rb samples had very similar covariance matrices, but the CPC procedure determined that Rb2 and Rb3 were “unrelated”, and that Rb1 and Rb2 had only one eigenvector in common. The few comparisons finding some eigenvectors in common did not correspond to particularly low measures of differentiation by the other methods. This shows that the CPC analysis did not consider the matrices’ properties in the same way.

## Discussion

The S statistics constitute sensitive tools for the detection of differences between covariance matrices. In the *Littorina* example used here, it was found that the local differentiation was clearly higher for the Su than for the Rb morph. This could be due to lower genetic connectivity for populations of the Su morph, and also to environmental differences between localities. However, the absence of such differences between the Rbs in the same localities would imply that any environmental differences would exist not between whole localities but only between the micro-habitats the Su snails use in the mid-shore. The differences between morphs, and specially those due to changes of orientation in the covariance matrices, were the largest in the analysis, and if not of the same size, they were of the same nature in all locations. Since shell morphology variation is adaptively important in *Littorina*[29], this suggests that these differences in covariance could be relevant for the evolutionary dynamics of the hybrid zone. They could simply result from the environmental differences between the two morphs’ microhabitats within the midshore, but even these environmental differences could have a genetic origin because individuals of the two morphs, even when living within extremely short distances of each other, tend to actively choose microhabitats with different physical characteristics [30].

The computer simulations shown in Figure 4 are limited to measure the power of the considered procedures to detect the consequences of changes in matrix orientation and shape. They show that such differences can be detected even when moderate in magnitude and when sample sizes are not too large, but cannot be taken as complete or definitive comparisons between the procedures. Matrices may differ in many relevant aspects, and different comparison procedures may have different aims and take different aspects into consideration. For example, while the S measures consider the differences in eigenstructure between two matrices, the RS procedure focuses on the related, but not equivalent problem of the differences between the evolutionary responses generated by these matrices. Comparisons would be even more difficult with procedures such as the set of evolvability measures proposed by Hansen and Houle [31], which consider the magnitude of the populations’ responses to different natural selection regimes.

Since the S statistics introduced here simply measure what proportion of variation exists in a given sample along the axis of variation defined by the eigenvectors in the compared sample, they are similar to the RS and T% ones in that they do not compare and are not dependent on the matrices’ sizes. They focus instead on the more interesting differences in matrix shape and orientation. In any case, S statistics-based comparisons could use raw covariance components instead of proportions as in the example shown, so that the results would depend on between-matrix size differences. However, in that case the S statistics would not be able to separate the effect of size from those of other sources of differentiation between matrices. Similarly, the basic version of the T method proposed in [19] reflects the differences in matrix size, as it is based on raw variance components instead of the proportions used by the T% statistic.

Calculating the amount of variance explained by a set of eigenvectors in a given dataset is straightforward in the case of datasets containing the phenotypic measures used to obtain **P** matrices. In the case of **G** matrices, the comparison would have to be based on additive genetic value estimates for individuals or families.

Since the proposed procedure is limited to two-sample comparisons, it cannot be used to make higher order analyses of the divergence among several populations (see [32]). However, and as shown in the *Litttorina* example, it can be useful for the study of evolutionary relevant situations such as hybrid zones. The S measures appear to be similar to the measure of distance between covariance matrices used by Mitteroecker and Bookstein [33], based on the calculation of relative eigenvalues, i.e., the eigenvalues of the product of one matrix premultiplied by the inverse of the second matrix, and therefore on expressing the variances and covariances of one sample relative to those of the other sample. I calculated the correlation of this measure with S1, S2 and S3 in a computer simulation considering the same cases as in Figure 4 and found that, while the measures were clearly related, they were not equivalent. For example, in the simulation of four variables and sample size 50, the correlation across replicates of the Mitteroecker and Bookstein measure with S1, S2 and S3 were: in the comparison of one population with itself, 0.206, 0.399 and 0.140 respectively; in the case of divergence in orientation, 0.396, 0.533 and 0.040^{NS}; in the case of divergence in matrix shape, 0.568, 0.281 and 0.554; and in the case of divergence in both orientation and shape, 0.454, 0.572 and 0.341 (all correlations, P < 0.001 except when indicated).

The *Littorina* example supports the view that CPCA might lead to misleading conclusions about the overall similarity between matrices [26], as pairs of matrices found very similar by other procedures were declared as “unrelated” by CPCA, and there was no clear correspondence between the two sets of results. The observed between-morphs reversal in importance of the second and third eigenvectors could play a role in this discrepancy. CPCA is based on a series of paired comparisons between eigenvectors of the same rank. Two matrices may share their axes of variation, but not the amount of variance in each axis. For example, the *ith* eigenvector of one matrix might be the same as the *i+1* eigenvector of the other, and the *i+1* of the first, the same as the *ith* of the second. In that case, the two matrices would have the same eigenvectors, but in a reverse order. A comparison between their *ith* eigenvectors would find them orthogonal, and this would also be the case for their *i+1* eigenvectors. Thus there may be considerable similarity between the two matrices, but this similarity is overlooked by the comparison procedure which finds the paired eigenvectors very different. The CPC software [34] enables users to compare the eigenvectors in any order, but this does not fix this particular limitation, as the order chosen for the two samples must be the same. The T and RS methods, based on matrices’ elements and product vectors, would provide a more balanced measure of similarity in this situation because the differences between these elements and these vectors would not depend on the existence of reversals in eigenvector order per se, but on the magnitude of the differences involved. However they don’t allow further analysis of the pattern of differentiation. The three S statistics are affected by different patterns of divergence, so that their joint use provided a deeper view of the differences between the two morphs’ **P** matrices. The S1 statistic is not dependent on the eigenvectors’ ordering per se because it is based on comparisons within eigenvector, i.e., on the difference between the amount of variance explained by one eigenvector from one sample in the original and reciprocal samples. These differences do not change with eigenvector order. But S2 changes when the order of eigenvectors in one of the samples is reversed (see formulas) because this would be considered as a change in matrix orientation. In case the reversal in eigenvectors’ importance was complete, so that there were no changes in overall shape, S3 would remain unaffected (see the second row in Figure 1). However, the reversal of the second and third eigenvectors between morphs cannot fully explain the disagreement between CPCA and the remaining methods because the results for S2 were not particularly similar to those of CPCA (Figure 5). This suggests that other aspects of covariance matrix structure might control the degree of agreement between different comparison procedures.

## Conclusions

The S-statistics procedure provides a simple and continuously-varying overall measure of differentiation that is distribution free and interpretable in terms of changes in matrix orientation and shape. In addition, it makes it easy to study the contribution of the different eigenvectors to the statistics values, which could provide further details on the nature of the differentiation, as was the case of the *Littorina* example used. This procedure could thus fill the gap between simpler statistics such as T% and RS, and more analytical methods like CPCA or Bayesian MCMC. The S-statistics procedure is not based on a formal analysis of matrices’ properties. Instead, it could serve for a simple and fast exploration of the magnitude and nature of the differentiation.

## Methods

### Compared procedures

The random skewers (RS) procedure was proposed by Cheverud [22]. Random selection vectors are generated by sampling selection coefficient values for each of the measured variables from a uniform distribution between 0 and 1. Then they are assigned positive or negative signs with 0.5 probabilities, assembled in selection vectors *β* and the sum of their squared coefficients made equal to one. Next, the responses corresponding to these selection vectors are calculated by multiplying them by each of the compared covariance matrices. If the compared matrices **A** and **B** are similar, the magnitude and direction of their responses to the same selection vector *β*_{
i
} will be similar. The correlation between the two response vectors **A** *β*_{
i
} and **B** *β*_{
i
} is calculated as

and the measure of similarity between matrices as the average correlation for all vectors.

In the T method [19] the differentiation between two matrices is measured as the sum of the absolute differences between the two matrices’ elements. In particular, I used the T% statistic, in which the sum is made relative to the average size of the elements in the two matrices [20]:

where *c* is the number of nonredundant elements in the matrices, M_{i1} and M_{i2} are such elements in the two matrices, and ${M}_{1}^{\u2015}$ and ${M}_{2}^{\u2015}$ their averages. The proportional nature of T% makes the comparison between matrices of different sizes easier. It is unreliable when there are both positive and negative elements in any of the two matrices [35], but this was not the case in the examples used in this work to compare the different methods’ performances.

Finally, I used the CPC (Common Principal Components) software of Phillips and Arnold [34] carrying out CPCA. I chose the “step-up” option, in which the likelihood of any model is tested against the likelihood of the next lowest model in the hierarchy, and used the Akaike Information Criterion to select the best description.

### Simulations

The simulations compared pairs of samples of individuals differing in the shape and orientation of their covariance matrices for the measured variables. All variables considered in the simulations had two normally-distributed components, one (*s*) specific for that variable and the other (*c*) common to all variables. The value for variable *i* measured on individual *j* in a given sample was generated as:

In each sample, matrix orientation was controlled by the relative contribution (fixed within sample) *k*_{i} of the common component to each variable’s value, and matrix shape, by the *s*_{i} variances. Four kinds of sample matrix comparisons were made: between samples taken at random from the same population, between samples from populations whose covariance matrices differed in orientation, whose matrices differed in shape, and whose matrices differed in both orientation and shape. One sample was taken at random from each of the two populations compared in each simulation case, and their covariance matrices and comparison statistics calculated. The observed value of each statistic was compared with the distribution obtained by comparing 50 pairs of resamples of the same size taken from the first sample, and with that obtained by comparing 50 pairs of resamples of the same size taken from the second sample. If the observed value was greater than these 100 resampled values, I concluded that the statistic found differences between the two samples’ matrices. This process was repeated 1000 times for each simulation case. I considered three sample sizes, 25, 50 and 100, and three numbers of variables, 2, 4 and 7 (the number of variables in the *Littorina* data). The particular changes in shape and orientation considered in each simulation case were chosen so that at least one of the three S statistics had nearly (but not exactly) 100% power to detect differences. The list of sample parameters used in the simulations is shown in Appendix 3.

*Littorina*data

I assayed the proposed matrix comparison method on six sets of shell morphology data from the two morphs (Rb and Su) of the marine snail *Littorina saxatilis*, taken in the Galician (NW Iberia) shores. The two morphs occupy different shore levels but form a hybrid zone in the midshore (see [36] for a detailed description of the hybrid zone). Seven shell measures were available (Figure 8, see [37] for more information), all of them taken on pure morph snails from the midshore (only a proportion of the individuals in this area are hybrids) of three locations (Centinela 42º04’N, 8º53’W; Corrubedo, 42º32’N, 9º20’W and Senin, .42º02’N, 8º53’W) and log transformed. The numbers of Rb and Su individuals from each location were 160 and 135, 150 and 112, and 123 and 111. I made three kinds of sample comparisons. First, I compared each sample with itself by generating pairs of pseudo-samples and their covariance matrices by resampling with replication in that sample’s dataset. Comparing samples with themselves provides a reference of maximum possible similarity, useful to interpret the similarities found between different samples. This is the “matrix repeatability” approach used in [26]. Second, I compared matrices from the same morph and different locations; and third, from different morphs. I obtained confidence distributions for the RS, T% , S1, S2 and S3 statistics by bootstrapping (999 replicates) the observations within each of the compared data sets, and obtained the statistics’ values for random pairs of the corresponding covariance matrix. This procedure was more demanding in the case of the RS statistic, because its observed values were already averages of correlations with randomized selection vectors (1000 vectors per value in this work).

## Appendix

### Appendix 1

The v_{ijk} values in Table 2 below, representing the variance explained by the *ith* eigenvector from the covariance matrix of sample *k* when applied to the data of sample *j*can be used to show that twice the sum of the squared differences across columns

Is the sum of the squared difference between diagonals:

and the squared difference between rows:

So that

and therefore

### Appendix 2

The covariance matrices used to draw plots in Figure 2 corresponded to pairs of variables *y*_{
i
} (*i* = 1, 2) defined as

where sqrt is the square root. Note that sqrt(*z*) is equivalent to *k* in the main text. The variances of *s* and *c* were:

The angles between eigenvector sets were determined by the *z* coefficients in each sample:

In this two-variables case, the variance explained by eigenvector *i* from sample *j* on sample *m* was calculated as

where **e**_{
ij
} and *e*_{
im
} are eigenvectors *i* from samples *j* and *m*, V_{im} is eigenvalue *i* from sample *m* and the apostrophe indicates transposition.

### Appendix 3

#### Summary of cases

The following tables show the values for the variances of variables *s* and *c* and the values of coefficient *z* in the expression:

used to generate the data for variable *i* and individual *j* in the different cases considered in the simulations -sqrt: square root; note that sqrt(*z*) is equivalent to *k* in the main text. Two variables

Four variables

Seven variables

*Detailed list of cases*

List of parameter sets used in every simulated case and resulting covariance matrices, eigenvectors and eigenvalues. The expected compositions of eigenvectors were obtained via eigenvector analyses applying R function *eigen* to random samples of size 10^{6}. Note that for four and seven variables cases it was not possible to obtain a constant set of eigenvector coefficients (beyond the first eigenvector) even for such large samples. In any case, The S statistics recognized their equivalence despite differences in eigenvectors’ coefficients (see the S3 and S2 values in the second row and third rows respectively of Figure 3).

In the two variables case we had:

Reference sample:

where sqrt is the square root and *s*_{1} and *s*_{2} had distributions N(0, 0.2), and *c*, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 1.8 and 0.2.

Compared sample with altered orientation:

where *s*_{1} and *s*_{2} had distributions N(0, 0.2), and c, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 1.8 and 0.2.

Compared sample with altered shape:

where s_{1} and s_{2} had distributions N(0, 0.5), and c, N(0, 0.5). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 1.5 and 0.5.

Compared sample with both orientation and shape altered:

where s_{1} and s_{2} had distributions N(0, 0.5), and c, N(0, 0.5). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 1.5 and 0.5.

The expected total variance in all two variable samples was = 2.

In the four variables case, we had:

Reference sample:

where *s*_{1,}*s*_{2,}*s*_{3} and *s*_{4} had distributions N(0, 0.2), and c, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 3.4, 0.2, 0.2, 0.2.

Compared sample with altered orientation:

where *s*_{1,}*s*_{2,}*s*_{3} and *s*_{4} had distributions N(0, 0.2), and c, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 3.4, 0.2, 0.2, 0.2.

Compared sample with altered shape:

where *s*_{1,}*s*_{2,}*s*_{3} and *s*_{4} had distributions N(0, 0.4), and c, N(0, 0.6). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 2.8, 0.4, 0.4, 0.4.

Compared sample with both orientation and shape altered:

where *s*_{1,}*s*_{2,}*s*_{3} and *s*_{4} had distributions N(0, 0.4), and c, N(0, 0.6). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 2.8, 0.4, 0.4, 0.4.

The expected total variance in all four variable samples was = 4.

In the seven variables case, we had:

Reference sample:

where *s*_{1,}*s*_{2,}*s*_{3,}*s*_{4,}*s*_{5,}*s*_{6,} and *s*_{7} had distributions N(0, 0.2), and c, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 5.8, 0.2, 0.2, 0.2, 0.2, 0.2, 0.2.

Compared sample with altered orientation:

where *s*_{1,}*s*_{2,}*s*_{3,}*s*_{4,}*s*_{5,}*s*_{6,} and *s*_{7} had distributions N(0, 0.2), and c, N(0, 0.8). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 5.8, 0.2, 0.2, 0.2, 0.2, 0.2, 0.2.

Compared sample with altered shape:

where s_{1,} s_{2,} s_{3,} s_{4,} s_{5,} s_{6,} and s_{7} had distributions N(0, 0.4), and c, N(0, 0.6). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 4.6, 0.4, 0.4, 0.4, 0.4, 0.4, 0.4.

Compared sample with both orientation and shape altered:

where s_{1,} s_{2,} s_{3,} s_{4,} s_{5,} s_{6,} and s_{7} had distributions N(0, 0.4), and c, N(0, 0.6). The expected covariance matrix was:

The expected eigenvectors had coefficients (columns):

and the expected eigenvalues were: 4.6, 0.4, 0.4, 0.4, 0.4, 0.4, 0.4.

The expected total variance in all seven variable samples was = 7.

## References

- 1.
Cheverud JM: A comparison of genetic and phenotypic correlations. Evolution. 1988, 42: 958-968. 10.2307/2408911.

- 2.
Ackermann RR, Cheverud JM: Morphological integration in primate evolution. Phenotypic Integration: Studying the Ecology and Evolution of Complex Phenotypes. Edited by: Pigliucci M, Preston K. 2004, New York: Oxford University Press, 302-319.

- 3.
Game ET, Caley J: The stability of P in coral reef fishes. Evolution. 2006, 60: 814-823.

- 4.
Kolbe JJ, Revell LJ, Szekely B, Brodie ED, Losos JB: Convergent evolution of phenotypic integration and its alignment with morphological diversification in Caribbean

*Anolis*ecomorphs. Evolution. 2011, 65: 3608-3624. 10.1111/j.1558-5646.2011.01416.x. - 5.
Jones AG, Arnold JS, Burger R: Evolution and stability of the G-matrix on a landscape with a moving optimum. Evolution. 2004, 58: 1639-1654.

- 6.
Arnold SJ, Bürger R, Hohenlohe PA, Ajie BC, Jones AG: Understanding the evolution and stability of the G-matrix. Evolution. 2008, 62: 2451-2461. 10.1111/j.1558-5646.2008.00472.x.

- 7.
Eroukhmanoff F, Svensson EI: Evolution and stability of the G-matrix during the colonization of a novel environment. J evol Biol. 2011, 24: 1363-1373. 10.1111/j.1420-9101.2011.02270.x.

- 8.
Colautti RI, Barrett SCH: Population divergence along lines of genetic variance and covariance in the invasive plant

*Lythrum salicaria*in eastern North America. Evolution. 2011, 65: 2514-2529. 10.1111/j.1558-5646.2011.01313.x. - 9.
Robinson MR, Wilson AJ, Pilkington JG, Clutton-Brock TH, Pemberton JM, Kruuk LEB: The impact of environmental heterogeneity on genetic architecture in a wild Population of Soay sheep. Genetics. 2009, 181: 1639-1648. 10.1534/genetics.108.086801.

- 10.
Monteiro LR, Nogueira MR: Adaptive radiations, ecological specialization, and the evolutionary integration of complex morphological structures. Evolution. 2010, 64: 724-744. 10.1111/j.1558-5646.2009.00857.x.

- 11.
Mallitt KL, Bonser SP, Hunt J: The plasticity of phenotypic integration in response to light and water availability in the pepper grass,

*Lepidium bonariense*. Evol Ecol. 2010, 24: 1321-1337. 10.1007/s10682-010-9373-6. - 12.
Barker BS, Phillips PC, Arnold SJ: A test of the conjecture that B-matrices are more stable than G-matrices. Evolution. 2010, 64: 2601-2613. 10.1111/j.1558-5646.2010.01023.x.

- 13.
Campbell DR, Weller SG, Sakai AK, Culley TM, Dang PN, Dunbar-Wallis AK: Genetic variation and covariation in floral allocation of two species of

*Schiedea*with contrasting levels of sexual dimorphism. Evolution. 2011, 65: 757-770. 10.1111/j.1558-5646.2010.01172.x. - 14.
Steppan SJ, Phillips PC, Houle D: Comparative quantitative genetics: evolution of the G matrix. Trends Ecol Evol. 2002, 17: 320-327. 10.1016/S0169-5347(02)02505-3.

- 15.
Hohenlohe PA, Arnold S: MIPoD: A hypothesis-testing framework for microevolutionary inference from patterns of divergence. Am Nat. 2008, 171: 366-385. 10.1086/527498.

- 16.
Ovaskainen O, Cano JM, Merilä J: A Bayesian framework for comparative quantitative genetics. Proc R Soc B. 2008, 275: 669-678. 10.1098/rspb.2007.0949.

- 17.
Kohn LAP, Atchley WR: How similar are genetic correlation structures? Data from mice and rats. Evolution. 1988, 42: 467-481. 10.2307/2409032.

- 18.
Atchley WR, Cowley DE, Vogl C, McLellan T: Evolutionary divergence, shape change, and genetic correlation structure in the rodent mandible. Syst Biol. 1995, 41: 196-221.

- 19.
Roff DA, Mousseau TA, Howard DJ: Variation in genetic architecture of calling song among populations of

*Allonemobius socius*,*A. fasciatus*, and a hybrid population: drift or selection?. Evolution. 1999, 53: 216-224. 10.2307/2640934. - 20.
Begin M, Roff DA: From micro- to macroevolution through quantitative genetic variation: positive evidence from field crickets. Evolution. 2004, 58: 2287-2304.

- 21.
Roff D: Comparing G matrices: a Manova approach. Evolution. 2002, 56: 1286-1291.

- 22.
Cheverud JM: Quantitative genetic analysis of cranial morphology in the cotton-top (

*Sagainus oedipas*) and saddle-back (*S. fuscicollis*) tamarins. J evol Biol. 1996, 9: 5-42. 10.1046/j.1420-9101.1996.9010005.x. - 23.
Calsbeek B, Goodnight CJ: Empirical comparison of G matrix test statistics: finding biologically relevant change. Evolution. 2009, 63: 2627-2635. 10.1111/j.1558-5646.2009.00735.x.

- 24.
Phillips PC, Arnold SJ: Hierarchical comparison of genetic variance-covariance matrices. I. Using the Flury hierarchy. Evolution. 1999, 53: 1506-1515. 10.2307/2640896.

- 25.
Flury B: Common Principal Components and Related Multivariate Models. 1988, New York: John Wiley

- 26.
Cheverud JM, Marroig G: Comparing covariance matrices: random skewers method compared to the common principal components model. Genet Mol Biol. 2007, 30: 461-469. 10.1590/S1415-47572007000300027.

- 27.
Houle D, Mezey J, Galpern P: Interpretation of the results of common principal components analyses. Evolution. 2002, 56: 433-440.

- 28.
Björklund M: Processes generating macroevolutionary patterns of morphological variation in birds: a simulation study. J evol Biol. 1994, 7: 727-742. 10.1046/j.1420-9101.1994.7060727.x.

- 29.
Cruz R, Vilas C, Mosquera J, García C: The close relationship between estimated divergent selection and observed differentiation supports the selective origin of a marine snail hybrid zone. J evol Biol. 2004, 17: 1221-1229. 10.1111/j.1420-9101.2004.00802.x.

- 30.
Carballo M, Caballero A, Rolán-Alvarez E: Habitat-dependent ecotype micro-distribution at the mid-shore in natural populations of

*Littorina saxatilis*. Hydrobiologia. 2005, 548: 307-311. 10.1007/s10750-005-1231-0. - 31.
Hansen TF, Houle D: Measuring and comparing evolvability and constraint in multivariate characters. J evol Biol. 2008, 21: 1201.1219-

- 32.
Hine E, Chenoweth S, Rundle HD, Blows MW: Characterizing the evolution of genetic variance using genetic covariance tensors. Phil Trans R Soc B. 2009, 364: 1567-1578. 10.1098/rstb.2008.0313.

- 33.
Mitteroecker M, Bookstein F: The ontogenetic trajectory of the phenotypic covariance matrix, with examples from craniofacial shape in rats and humans. Evolution. 2009, 63: 727-737. 10.1111/j.1558-5646.2008.00587.x.

- 34.
Phillips P: CPC – Common principal Components Analysis Program. http://pages.uoregon.edu/pphil/programs/cpc/cpc.htm,

- 35.
Steppan SJ: Phylogenetic analysis of phenotypic covariance structure. II. Reconstructing matrix evolution. Evolution. 1997, 51: 587-594. 10.2307/2411130.

- 36.
Johannesson K, Johannesson B, Rolán-Alvarez E: Morphological differentiation and genetic cohesiveness over a micro environmental gradient in the marine snail

*Littorina saxatilis*. Evolution. 1993, 47: 1770-1787. 10.2307/2410220. - 37.
Cruz R, Rolán-Álvarez E, García C: Sexual selection on phenotypic traits in a hybrid zone of

*Littorina saxatilis*(Olivi). J evol Biol. 2001, 14: 773-785.

## Acknowledgements

I thank Raquel Cruz, Javier Mosquera and Carlos Vilas for allowing me to use the data from our previous experiments, which had been funded by Spain’s DGICYT grant PB94-0649, and Paul Hohenlohe and David Houle for helpful criticism of the manuscript. This work was funded by Ministerio de Ciencia y Tecnología (CGL2009-13278-C02),) and Fondos Feder.

## Author information

## Additional information

### Competing interests

The author declares that he has no competing interests.

## Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

## Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

## About this article

#### Received

#### Accepted

#### Published

#### DOI

### Keywords

- Eigenvectors
- Principal component analysis
- Littorina Saxatilis
- Matrix orientation
- Matrix shape
- Hybrid zone