Siegel–Tukey Test
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, the Siegel–Tukey test, named after
Sidney Siegel Sidney Siegel (4 January 1916 in New York City – 29 November 1961) was an American psychologist who became especially well known for his work in popularising non-parametric statistics for use in the behavioural sciences. He was a co-developer o ...
and
John Tukey John Wilder Tukey (; June 16, 1915 – July 26, 2000) was an American mathematician and statistician, best known for the development of the fast Fourier Transform (FFT) algorithm and box plot. The Tukey range test, the Tukey lambda distributi ...
, is a
non-parametric test Nonparametric statistics is the branch of statistics that is not based solely on parametrized families of probability distributions (common examples of parameters are the mean and variance). Nonparametric statistics is based on either being dist ...
which may be applied to data measured at least on an ordinal scale. It tests for differences in scale between two groups. The test is used to determine if one of two groups of data tends to have more widely dispersed values than the other. In other words, the test determines whether one of the two groups tends to move, sometimes to the right, sometimes to the left, but away from the center (of the ordinal scale). The test was published in 1960 by
Sidney Siegel Sidney Siegel (4 January 1916 in New York City – 29 November 1961) was an American psychologist who became especially well known for his work in popularising non-parametric statistics for use in the behavioural sciences. He was a co-developer o ...
and John Wilder Tukey in the '' Journal of the American Statistical Association'', in the article "A Nonparametric Sum of Ranks Procedure for Relative Spread in Unpaired Samples."


Principle

The principle is based on the following idea: Suppose there are two groups A and B with ''n'' observations for the first group and ''m'' observations for the second (so there are ''N'' = ''n'' + ''m'' total observations). If all ''N'' observations are arranged in ascending order, it can be expected that the values of the two groups will be mixed or sorted randomly, if there are no differences between the two groups (following the null hypothesis H0). This would mean that among the ranks of extreme (high and low) scores, there would be similar values from Group A and Group B. If, say, Group A were more inclined to extreme values (the
alternative hypothesis In statistical hypothesis testing, the alternative hypothesis is one of the proposed proposition in the hypothesis test. In general the goal of hypothesis test is to demonstrate that in the given condition, there is sufficient evidence supporting ...
H1), then there will be a higher proportion of observations from group A with low or high values, and a reduced proportion of values at the center. :* Hypothesis H0: σ2A = σ2B & MeA = MeB (where σ2 and Me are the variance and the median, respectively) :* Hypothesis H1: σ2A > σ2B


Method

Two groups, A and B, produce the following values (already sorted in ascending order): : A: 33 62 84 85 88 93 97     B: 4 16 48 51 66 98 By combining the groups, a group of 13 entries is obtained. The ranking is done by alternate extremes (rank 1 is lowest, 2 and 3 are the two highest, 4 and 5 are the two next lowest, etc.). The sum of the ranks within each W group: : ''W''A = 5 + 12 + 11 + 10 + 7 + 6 + 3 = 54 : ''W''B = 1 + 4 + 8 + 9 + 13 + 2 = 37 If the null hypothesis is true, it is expected that the average ranks of the two groups will be similar. If one of the two groups is more dispersed its ranks will be lower, as extreme values receive lower ranks, while the other group will receive more of the high scores assigned to the center. To test the difference between groups for significance a
Wilcoxon rank sum test Wilcoxon is a surname, and may refer to: * Charles Wilcoxon, drum educator * Henry Wilcoxon, an actor * Frank Wilcoxon, chemist and statistician, inventor of two non-parametric tests for statistical significance: ** The Wilcoxon signed-rank test (a ...
is used, which also justifies the notation WA and WB in calculating the rank sums. From the rank sums the U statistics are calculated by subtracting off the minimum possible score, ''n''(''n'' + 1)/2 for each group:Lehmann, Erich L., ''Nonparametrics: Statistical Methods Based on Ranks'', Springer, 2006, pp. 9, 11–12. : ''U''A = 54 − 7(8)/2 = 26 : ''U''B = 37 − 6(7)/2 = 16 According to H_0 the minimum of these two values is distributed according to a Wilcoxon rank-sum distribution with parameters given by the two group sizes: ::: \min(U_A,U_B) \sim \text(m,n) \! Which allows the calculation of a p-value for this test according to the following formula: :::p = \Pr\left \le \min(U_A,U_B) \right\,\! :::X \sim \text(m,n)\,\! a table of the Wilcoxon rank-sum distribution can be used to find the statistical significance of the results (see
Mann–Whitney_U_test In statistics, the Mann–Whitney ''U'' test (also called the Mann–Whitney–Wilcoxon (MWW/MWU), Wilcoxon rank-sum test, or Wilcoxon–Mann–Whitney test) is a nonparametric test of the null hypothesis that, for randomly selected values ''X'' ...
for more explanations on these tables). For the example data, with groups of sizes m=6 and n=7 the p-value is: :::p=\Pr\left \le 16 \right= 0.2669.\,\! indicating little or no reason to reject the null hypothesis that the dispersion of the two groups is the same.


See also

* Non-parametric statistics *
Statistical hypothesis testing A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters. ...


References


External links


an R implementation of Siegel-Tukey test
{{DEFAULTSORT:Siegel-Tukey test Statistical tests Nonparametric statistics