Kruskal–Wallis test

The Kruskal–Wallis test by ranks, Kruskal–Wallis test[1] (named after William Kruskal and W. Allen Wallis), or one-way ANOVA on ranks[1] is a non-parametric method for testing whether samples originate from the same distribution.[2][3][4] It is used for comparing two or more independent samples of equal or different sample sizes. It extends the Mann–Whitney U test, which is used for comparing only two groups. The parametric equivalent of the Kruskal–Wallis test is the one-way analysis of variance (ANOVA).

Difference between ANOVA and Kruskal-Wallis test with ranks

A significant Kruskal–Wallis test indicates that at least one sample stochastically dominates one other sample. The test does not identify where this stochastic dominance occurs or for how many pairs of groups stochastic dominance obtains. For analyzing the specific sample pairs for stochastic dominance, Dunn's test,[5] pairwise Mann–Whitney tests with Bonferroni correction,[6] or the more powerful but less well known Conover–Iman test[6] are sometimes used.

It is supposed that the treatments significantly affect the response level and then there is an order among the treatments: one tends to give the lowest response, another gives the next lowest response is second, and so forth.[7] Since it is a nonparametric method, the Kruskal–Wallis test does not assume a normal distribution of the residuals, unlike the analogous one-way analysis of variance. If the researcher can make the assumptions of an identically shaped and scaled distribution for all groups, except for any difference in medians, then the null hypothesis is that the medians of all groups are equal, and the alternative hypothesis is that at least one population median of one group is different from the population median of at least one other group. Otherwise, it is impossible to say, whether the rejection of the null hypothesis comes from the shift in locations or group dispersions. This is the same issue that happens also with the Mann-Whitney test.[8][9][10] If the data contains potential outliers, if the population distributions have heavy tails, or if the population distributions are significantly skewed, the Kruskal-Wallis test is more powerful at detecting differences among treatments than ANOVA F-test. On the other hand, if the population distributions are normal or are light-tailed and symmetric, then ANOVA F-test will generally have greater power which is the probability of rejecting the null hypothesis when it indeed should be rejected.[11][12]

Method

An illustration of how to assign any tied values the average of the rank
  1. Rank all data from all groups together; i.e., rank the data from 1 to N ignoring group membership. Assign any tied values the average of the ranks they would have received had they not been tied.
  2. The test statistic is given by
    where
    • is the total number of observations across all groups
    • is the number of groups
    • is the number of observations in group
    • is the rank (among all observations) of observation from group
    • is the average rank of all observations in group
    • is the average of all the .
  3. If the data contain no ties the denominator of the expression for is exactly and . Thus

    The last formula only contains the squares of the average ranks.
  4. A correction for ties if using the short-cut formula described in the previous point can be made by dividing by , where G is the number of groupings of different tied ranks, and ti is the number of tied values within group i that are tied at a particular value. This correction usually makes little difference in the value of H unless there are a large number of ties.
  5. When performing multiple sample comparisons, the type I error tends to become inflated. Therefore, Bonferroni procedure is used to adjust the significance level, that is, , where is the adjusted significance level, is the initial significance level, and is the number of the contrasts.[13]
  6. Finally, the decision to reject or not the null hypothesis is made by comparing to a critical value obtained from a table or a software for a given significance or alpha level. If is bigger than , the null hypothesis is rejected. If possible (no ties, sample not too big) one should compare to the critical value obtained from the exact distribution of . Otherwise, the distribution of H can be approximated by a chi-squared distribution with g-1 degrees of freedom. If some values are small (i.e., less than 5) the exact probability distribution of can be quite different from this chi-squared distribution. If a table of the chi-squared probability distribution is available, the critical value of chi-squared, , can be found by entering the table at g − 1 degrees of freedom and looking under the desired significance or alpha level.[14]
  7. If the statistic is not significant, then there is no evidence of stochastic dominance between the samples. However, if the test is significant then at least one sample stochastically dominates another sample. Therefore, a researcher might use sample contrasts between individual sample pairs, or post hoc tests using Dunn's test, which (1) properly employs the same rankings as the Kruskal–Wallis test, and (2) properly employs the pooled variance implied by the null hypothesis of the Kruskal–Wallis test in order to determine which of the sample pairs are significantly different.[5] When performing multiple sample contrasts or tests, the Type I error rate tends to become inflated, raising concerns about multiple comparisons.

Exact probability tables

A large amount of computing resources is required to compute exact probabilities for the Kruskal–Wallis test. Existing software only provides exact probabilities for sample sizes of less than about 30 participants. These software programs rely on the asymptotic approximation for larger sample sizes.

Exact probability values for larger sample sizes are available. Spurrier (2003) published exact probability tables for samples as large as 45 participants.[15] Meyer and Seaman (2006) produced exact probability distributions for samples as large as 105 participants.[16]

Exact distribution of H

Choi et al.[17]made a review of two methods that had been developed to compute the exact distribution of , proposed a new one, and compared the exact distribution to its chi-squared approximation.

Example

Test for differences in ozone levels by month

The following example uses data from Chambers et al.[18]on daily readings of ozone for May 1 to September 30, 1973, in New York City. The data are in the R data set airquality, and the analysis is included in the documentation for the R function kruskal.test. Boxplots of ozone values by month are shown in the figure.

The Kruskal-Wallis test finds a significant difference (p = 6.901e-06) indicating that ozone differs among the 5 months.

kruskal.test(Ozone ~ Month, data = airquality)Kruskal-Wallis rank sum testdata:  Ozone by MonthKruskal-Wallis chi-squared = 29.267, df = 4, p-value = 6.901e-06

To determine which months differ, post-hoc tests may be performed using a Wilcoxon test for each pair of months, with a Bonferroni (or other) correction for multiple hypothesis testing.

pairwise.wilcox.test(airquality$Ozone, airquality$Month, p.adjust.method = "bonferroni")Pairwise comparisons using Wilcoxon rank sum testdata:  airquality$Ozone and airquality$Month  5      6      7      8     6 1.0000 -      -      -     7 0.0003 0.1414 -      -     8 0.0012 0.2591 1.0000 -     9 1.0000 1.0000 0.0074 0.0325P value adjustment method: bonferroni

The post-hoc tests indicate that, after Bonferroni correction for multiple testing, the following differences are significant (adjusted p < 0.05).

  • Month 5 vs Months 7 and 8
  • Month 9 vs Months 7 and 8

Implementation

The Kruskal-Wallis test can be implemented in many programming tools and languages.

  • Mathematica implements the test as LocationEquivalenceTest.[19]
  • MATLAB's Statistics Toolbox has kruskalwallis to compute the p-value for a hypothesis test and display ANOVA table.[20]
  • SAS has the "NPAR1WAY" procedure for the test.[21]
  • SPSS implements the test with the "Nonparametric Tests" procedure.[22]
  • Minitab has the implement in the "Nonparametrics" option.[23]
  • In Python's SciPy package, the function scipy.stats.kruskal can return the test result and p-value.[24]
  • R base-package has an implement of this test using kruskal.test.[25]
  • Java has the implement provided by provided by Apache Commons.[26]
  • In Julia, the package HypothesisTests.jl has the function KruskalWallisTest(groups::AbstractVector{<:Real}...) to compute the p-value.[27]

See also

References

Further reading

External links