Abstract

This paper studies correction for chance for association measures for continuous variables. The set of linear transformations of Pearson’s product-moment correlation is used as the domain of the correction for chance function. Examples of measures in this set are Tucker’s congruence coefficient, Jobson’s coefficient, and Pearson’s correlation. An equivalence relation on the set of linear transformations is defined. The fixed points of the correction for chance function are characterized. It is shown that each linear transformation is mapped to the fixed point in its equivalence class.

1. Introduction

In various subfields of statistics, association measures are commonly used to express the strength of a relationship between two variables in a number. Individual measures are used for summarizing parts of a research study, while matrices of association measures can be used as input for multivariate data analysis techniques like regression analysis and component analysis [1, 2]. In this paper, we study measures that quantify the association between two continuous variables. Examples of these measures are Tucker’s congruence coefficient, Jobson’s coefficient, and Pearson’s product-moment correlation. The latter coefficient is commonly used as a measure of the linear dependence between two continuous variables.

Pearson’s correlation is a measure that has multiple interesting interpretations [3]. For example, the squared correlation may be interpreted as the percentage of variance that the two variables have in common. However, for many other association measures, only the extreme values have a clear interpretation. Many association measures have maximum value unity, indicating perfect association. However, a zero value is usually attained in different situations. For example, Pearson’s correlation and Jobson’s coefficient have zero value under statistical independence, while Tucker’s congruence coefficient has zero value if the inner product of the two variables equals zero.

It may be desirable that an association measure has zero value if two variables are statistically independent [46]. If a measure does not have zero value under statistical independence, a standard approach in the literature is to correct it for association due to chance [79]. After correction for chance, a measure has a form where is the value of measure under chance and 1 in the denominator of (1) is the maximum value of . In this paper, we only consider measures with maximum value unity. Furthermore, we assume that to avoid the indeterminate case . Transformation (1) sets the value of a measure at zero under independence, while leaving the maximum value at unity.

Transformation (1) has been almost exclusively studied in relation to association measures for binary variables. Measures for binary variables are also called coefficients for tables. Several authors showed that certain measures become equivalent after correction (1). For example, Fleiss [8], Zegers [6], Albatineh et al. [7], and Warrens [9] showed that measures by Dice, Hamann, and Rand become Cohen’s kappa after correction for chance. Furthermore, Warrens [10] characterized all measures that are transformed into weighted kappa [10, 11] by formula (1).

The abovementioned results provide new interpretations of various chance-corrected measures. Furthermore, if different measures coincide after correction for chance, the task of choosing the best chance-corrected measure for the data-analytic task at hand becomes easier, since there are less measures to choose from. Moreover, Warrens [9, 10, 12] studied transformation (1) as a mathematical function. These latter studies revealed how different association measures from the literature are related and also provided new ways to interpret chance-corrected measures.

In this paper, we study correction for chance for association measures for continuous variables. Several results for measures are generalized to the more complex case of continuous variables. Furthermore, various new results illustrate how certain measures are related and provide new ways for interpreting them. For example, we show that Pearson’s correlation may be interpreted as a chance-corrected measure, which is a new result.

The paper is organized as follows. In Section 2, we introduce the notation and definitions of the association measures, and we specify the domain of the correction for chance function. In Section 3, we briefly discuss statistical inference based on measures in this domain. In Section 4, we define the correction for chance function and present some of its properties. Some of the results presented in this section generalize properties presented in Warrens [9] for measures for tables. In Section 5, we consider several families of association measures that belong to the same equivalence classes. In Section 6, we characterize the fixed points of the function in (1). Finally, Section 7 contains a discussion.

2. A Set of Association Measures

Since many association measures in the literature have been proposed as sample statistics [13, 14], we will study the measures here as sample statistics. For certain measures, the population equivalents can be obtained by replacing sample variances and covariances by their corresponding population parameters, degenerate cases excluded.

Suppose we have scores of objects (subjects, individuals) on two variables and . An individual score is denoted by for . Let the summation be short for . The mean score of is given by the unbiased sample variance is given by and the mean squared value of is given by The sample statistics for are defined analogously. The unbiased sample covariance of and is given by Association measures can be defined as functions from the set of pairs of variables of length into the real numbers. The set will be used as the domain. For many measures from the literature, the codomain is either the interval or the interval . An association measure is defined as a function , and the set of all such measures is denoted by . A measure that belongs to is Pearson’s correlation coefficient given by Pearson’s correlation is a commonly used measure of linear dependence between two variables.

In the following, we are interested in a particular subset of . Kendall and Stuart [15, p. 492] discuss the permutation distribution of . The exact distribution of can be obtained by considering all permutations of the scores on , while keeping the scores on fixed. Under permutation of the scores on , only is a random variable. For each permutation, we may calculate . The expectation can then be defined as the average of all the values of . Kendall and Stuart [15, p. 492] show that Since only is a random variable under permutation, we can write as a linear transformation of ; that is, , where are functions that are not affected by permutation of the scores on or . More generally, we may consider the set given by The elements of are linear transformations of , and and are functions that remain unchanged under permutations of the scores on or . Furthermore, 1 is the maximum value of the measures in . Since Pearson’s correlation belongs to , the set is nonempty. Examples of other association measures that belong to are presented in Section 5.

Instead of considering , we may also consider the set given by Measures in are linear transformations of , and and are functions that are unaffected by permutations of the scores on or . Lemma 1 shows that and are equivalent.

Lemma 1. .

Proof. Consider (7). We can write as , where Hence, and it follows that . Furthermore, we can write as , where and are presented in (9). Hence, and it follows that . Thus, .

Lemma 1 shows that linearity in is equivalent to linearity in . The proof of Lemma 1 illustrates that any element of can be used as a generator of . For example, we have . The set is used as the domain of the correction for chance formula in the next section. A similar family of linear transformations has been considered in [5, 7, 9, 10, 16].

3. Statistical Inference

In this paper, we mainly present various algebraic and analytic properties of the broad family of association measures in (10) and (11). In this section, we briefly discuss statistical inference based on members of this family. In the framework presented here, variables are considered random if they are not invariant under permutation [14, 17]. Therefore, the quantity is the only random variable in (10). A straightforward approach to performing hypothesis tests or constructing confidence intervals for a member of the family in (10) could be based on a permutation test with respect to the quantity . Furthermore, confidence intervals for a member of the family in (10) may be obtained using the bootstrap.

It should be noted that the assumption that variables are random only if they are not invariant under permutation is unusual for Pearson’s correlation coefficient. As pointed out by a reviewer, with Pearson’s correlation, it is common to assume that and follow a bivariate normal distribution. In this case, the quantities , , , , , and are all random variables.

4. Correction for Chance

Equation (1) presents the formula of a measure after correction for chance. For measures in , the correction for chance function is defined as Lemma 2 shows that, for measures in , formula (13) becomes formula (14).

Lemma 2. Let with . One has

Proof. Using identity (8), we have for measures in the identity Using identity (15) in formula (13) and dividing all terms in the result by , we obtain or, equivalently, formula (14).

Formula (14) shows that elements of coincide after correction for chance if they have the same ratio This property of elements in will be used repeatedly in the following. A similar property was found in [7] for validation indices in cluster analysis. Furthermore, for association measures for tables, a similar property was found in [9].

The function in (13) is a map from to if is closed under . Lemma 3 shows that this is the case.

Lemma 3. is closed under .

Proof. Let with . The formula for is presented in (16). We can write as , where Hence, .

Let , let with , and consider the linear transformation . Since we require that all association measures have maximum value 1, we have the restriction . Hence, is also an element in . Lemma 4 shows that and coincide after correction for chance. A proof of Lemma 4 for the special case of measures can be found in Warrens [9].

Lemma 4. Let , let , and consider . One has .

Proof. Since , there exist and such that . Hence, and we can write as , where Using these quantities, ratio (17) is equal to Since ratio (17) is identical for and , it follows that .

Suppose we have several association measures that coincide after correction for chance. Lemma 5 shows that the arithmetic mean of the association measures becomes identical to the measures after correction for chance. A proof of Lemma 5 for the special case of measures can be found in Warrens [9].

Lemma 5. Let with for such that The arithmetic mean coincides with after correction for chance.

Proof. We can write as , where Using these quantities, ratio (17) is equal to The right-hand side of this equation shows that is a weighted average of using weights . Since are all identical, is identical to . Hence, .

5. Equivalence Classes

Ratio (17) divides the association measures in into equivalence classes with respect to function in (14). We have the following definition. Two measures are said to be equivalent with respect to (14), denoted by , if they have the same ratio (17). It can be shown that is an equivalence relation on ; that is, is reflexive, symmetric, and transitive.

The equivalence relation divides the elements of into equivalence classes, one class for each value of (17). We consider some examples. Let be a function that remains unchanged under permutations of the scores on or , and consider the family If ,  (24) is identical to the coefficient of identity [6, 14, 17]. If , (24) is identical to Jobson’s coefficient [18], which is identical to the chance-corrected identity coefficient [6, 17]. We can write (24) as , where Using and in (26), ratio (17) is equal to Hence, all special cases of (24), regardless of the specification of , belong to the equivalence class that is characterized by ratio (27). Furthermore, all special cases of (24) coincide after correction for chance.

Next, consider the family If , (28) is identical to Tucker’s congruence coefficient [14, 19], also called the coefficient of proportionality [6, 14]. If , (28) is identical to the chance-corrected proportionality coefficient [6, 17]. We can write (28) as , where Using and in (29), ratio (17) is equal to Hence, all special cases of (28) belong to the equivalence class that is characterized by ratio (30). Furthermore, all special cases of (28) coincide after correction for chance.

Next, consider the family If , (31) is identical to Pearson’s correlation in (7). We can write (31) as , where Using and in (32), ratio (17) is equal to Hence, all special cases of (31) belong to the equivalence class that is characterized by ratio (33). Furthermore, all special cases of (31) coincide after correction for chance.

The examples in (24), (28), and (31) illustrate that many measures from the literature belong to the set . Furthermore, the examples show that the equivalence classes have infinitely many elements. Since multiple measures coincide after correction for chance, the function in (14) is not injective. Since is not injective, it is not invertible. Furthermore, for , we have, in general, . Thus, in general, is not a linear map. Lemmas 4 and 5 show that is a linear map if the domain of is restricted to an equivalence class for a specific value of ratio (17). All association measures in this equivalence class and their linear combinations are mapped onto a measure of the form (14). In the next section, we characterize these so-called fixed points.

6. Fixed Points

A measure is called a fixed point of if holds. Using and in (18), we have Hence, . In other words, the function is idempotent. An idempotent function has at least one fixed point. Suppose is not a fixed point and let . Since is idempotent, we have . Hence, elements of that are not fixed points are mapped to fixed points.

In the remainder of this section, we characterize the fixed points of . Let be a function that remains unchanged under permutations of the scores on or , and consider the set given by The elements in are association measures of the form , where is the sample covariance of variables and . Lemma 6 shows that is the set of fixed points of .

Lemma 6. The set is the set of fixed points of .

Proof. () Let . Then, for some . Using in (16), we obtain Hence, , and it follows that is a fixed point.
() Let with be a fixed point. Then, , or equivalently Equating -parts and “not”--parts on both sides of the equality, we obtain the identities Using the identity , we can write as Hence, . This completes the proof.

Since points that are not fixed points of are mapped to fixed points, we have ; that is, the images of are the fixed points in . It follows from Lemma 6 that Pearson’s correlation in (7) and Jobson’s coefficient in (25) are two fixed points of . All measures that belong to the equivalence class characterized by ratio (33) are mapped to Pearson’s correlation. All measures that belong to the equivalence class characterized by ratio (27) are mapped to Jobson’s coefficient.

7. Discussion

In this paper, we studied correction for chance as a mathematical function in the context of association measures for continuous variables. The function is not injective and in general not a linear map. All association measures that are not fixed points of the function are mapped to fixed points. Notable fixed points are Jobson’s coefficient and Pearson’s correlation. Both coefficients are fixed points of distinct equivalence classes of association measures. Characteristics of association measures in both classes were presented.

The results show that both Jobson’s coefficient and Pearson’s correlation may be interpreted as chance-corrected measures. For Pearson’s correlation, this interpretation is new. Thirteen other ways to look at Pearson’s correlation are discussed in Rodgers and Nicewander [3]. What is also new is that there are infinitely many measures that become Jobson’s coefficient and infinitely many coefficients that coincide with Pearson’s correlation after correction for chance. A data-analytic choice between Jobson’s coefficient and Pearson’s correlation could be based on the metric scales discussed in Zegers [6, 17].

Conflict of Interests

The author declares that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

The author thanks Professor Heinz Holling for the helpful comments and valuable suggestions on an earlier version of this paper.