Applied Univariate, Bivariate, and Multivariate Statistics. Daniel J. Denis
Actual scores on the favorability measure are in parentheses.
To compute Spearman's rs in R the “long way,” we generate two vectors that contain the respective rankings:
> bill <- c(5, 1, 3, 4, 2) > mary <- c(5, 3, 1, 4, 2)
Because the data are already in the form of ranks, both Pearson r and Spearman rho will agree:
> cor(bill, mary) [1] 0.6 > cor(bill, mary, method = “spearman”) > 0.6
Note that by default, R returns the Pearson correlation coefficient. One has to specify method = “spearman”
to get rs. Consider now what happens when we correlate, instead of rankings, the actual subjective favorability scores corresponding to the respective ranks. When we plot the favorability data, we obtain:
> bill.sub <- c(2.1, 7.6, 8.4, 9.5, 10.0) > mary.sub <- c(7.6, 8.5, 9.0, 9.6, 9.7) > plot(mary.sub, bill.sub)
Note that though the relationship is not perfectly linear, each increase in Bill's subjective score is nonetheless associated with an increase in Mary's subjective score. When we compute Pearson's r on this data, we obtain:
> cor(bill.sub, mary.sub) [1] 0.9551578
However, when we compute rs, we get:
> cor(bill.sub, mary.sub, method = "spearman") [1] 1
Spearman's rs is equal to 1.0 because the rankings of movie preferences are perfectly monotonically increasing (i.e., for each increase in movie preference along the abscissa corresponds an increase in movie preference along the ordinate). In the case of Pearson's, the correlation is less than 1.0 because r captures the linear relationship among variables and not simply a monotonically increasing one. Hence, a high magnitude coefficient for Spearman's essentially tells us that two variables are “moving together,” but it does not necessarily imply the relationship is a linear one. A similar test that measures rank correlation is that of Kendall's rank‐order correlation. See Siegel and Castellan (1988, p. 245) for details.
2.20 STUDENT'S t DISTRIBUTION
The density for Student's t is given by (Shao, 2003):
where Γ is the gamma function and v are degrees of freedom. For small degrees of freedom v, the t distribution is quite distinct from the standard normal. However, as degrees of freedom increase, the t distribution converges to that of a normal density (Figure 2.11). That is, in the limit, f(t) → f(z), or a bit more formally,
The fact that t converges to z for large degrees of freedom but is quite distinct from z for small degrees of freedom is one reason why t distributions are often used for small sample problems. When sample size is large, and so consequently are degrees of freedom, whether one treats a random variable as t or z will make little difference in terms of computed p‐values and decisions on respective null hypotheses. This is a direct consequence of the convergence of the two distributions for large degrees of freedom. For a historical overview of how t‐distributions came to be, consult Zabell (2008).
2.20.1 t‐Tests for One Sample
When we perform hypothesis testing using the z distribution, we assume we have knowledge of the population variance σ2. Having direct knowledge of σ2 is the most ideal and preferable of circumstances. When we know σ2, we can compute the standard error of the mean directly as
Figure 2.11 Student's t versus normal densities for 3 (left), 10 (middle), and 50 (right) degrees of freedom. As degrees of freedom increase, the limiting form of the t distribution is the z distribution.
Recall that the form of the one‐sample z test for the mean is given by
where the numerator
In most research contexts, from simple to complex, we usually do not have direct knowledge of σ2. When we do not have knowledge of it, we use the next best thing, an estimate of it. We can obtain an unbiased estimate of σ2 by computing s2 on our sample. When we do so, however, and use s2 in place of σ2, we can no longer pretend to “know” the standard error of the mean. Rather, we must concede that all we are able to do is estimate it. Our estimate of the standard error of the mean is thus given by:
When we use s2 (where
was found to be distributed as a t statistic on n − 1 degrees of freedom. Again, the t distribution is most useful for when sample sizes are rather small. For larger samples, as mentioned, the t distribution converges to that of the z distribution. If you are using rather large samples, say approximately 100 or more, whether you evaluate your null hypothesis using a z or t distribution will not matter much, because the critical values for z and t for such degrees of freedom (99 for the one‐sample case) will be relatively alike, that practically at least, the two test statistics can be considered more or less equal. For even larger samples, the convergence is that much more fine‐tuned.
The concept of convergence between z and t can be easily illustrated by inspecting the variance of the t distribution. Unlike the z distribution where the variance is set at 1.0 as a constant, the variance of the t distribution is defined as: