Classification of m hypothesis tests
The following table gives a number of errors committed when testing| Null hypothesis is True (H0) | Alternative hypothesis is True (H1) | Total | |
|---|---|---|---|
| Declared significant | |||
| Declared non-significant | |||
| Total |
is the total number hypotheses tested
is the number of true null hypotheses
is the number of true alternative hypotheses
is the number of false positives (Type I error) (also called "false discoveries")
is the number of true positives (also called "true discoveries")
is the number of false negatives (Type II error)
is the number of true negatives
is the number of rejected null hypotheses (also called "discoveries")
- In
hypothesis tests of which
are true null hypotheses,
is an observable random variable, and
,
,
, and
are unobservable random variables.
False discovery rate (FDR)
Based on previous definitions we can defineAnd one wants to keep this value below a threshold
待看資料:http://brainder.org/2011/09/05/fdr-corrected-fdr-adjusted-p-values/
Familywise error rate (FWER)
The FWER is the probability of making even one type I error In the family,A procedure controls the FWER in the weak sense if the FWER control at level
A procedure controls the FWER in the strong sense if the FWER control at level
False discovery rate (FDR):
FDR procedures are designed to control the expected proportion of incorrectly rejected null hypotheses ("false discoveries").[1] FDR controlling procedures exert a less stringent control over false discovery compared to familywise error rate (FWER) procedures (such as the Bonferroni correction), which seek to reduce the probability of even one false discovery, as opposed to the expected proportion of false discoveries. Thus FDR procedures have greater power at the cost of increased rates of type I errors, i.e., rejecting the null hypothesis of no effect when it should fail to be rejected.
Post-hoc testing of ANOVAs
Multiple comparison procedures are commonly used in an analysis of variance after obtaining a significant omnibus test result, like the ANOVA F-test. The significant ANOVA result suggests rejecting the global null hypothesis H0 that the means are the same across the groups being compared. Multiple comparison procedures are then used to determine which means differ. In a one-way ANOVA involving K group means, there are K(K − 1)/2 pairwise comparisons.A number of methods have been proposed for this problem, some of which are:
- Single-step procedures
- Tukey–Kramer method (Tukey's HSD) (1951)
- Scheffe method (1953)
- Rodger's method (precludes type 1 error rate inflation, using a decision-based error rate)
- Multi-step procedures based on Studentized range statistic
- Duncan's new multiple range test (1955)
- The Nemenyi test is similar to Tukey's range test in ANOVA.
- The Bonferroni–Dunn test allows comparisons, controlling the familywise error rate.[vague]
- Student Newman-Keuls post-hoc analysis
- Dunnett's test (1955) for comparison of number of treatments to a single control group.
For example,if the variances of the groups being compared are similar, the Tukey–Kramer method is generally viewed as performing optimally or near-optimally in a broad variety of circumstances.[8] The situation where the variance of the groups being compared differ is more complex, and different methods perform well in different circumstances.
The Kruskal–Wallis test is the non-parametric alternative to ANOVA. Multiple comparisons can be done using pairwise comparisons (for example using Wilcoxon rank sum tests) and using a correction to determine if the post-hoc tests are significant (for example a Bonferroni correction).
Holm–Bonferroni method
In statistics, the Holm–Bonferroni method [1] is a method used to counteract the problem of multiple comparisons. It is intended to control the Familywise error rate and offers a simple test uniformly more powerful than the Bonferroni correction. It is one of the earliest usage of stepwise algorithms in simultaneous inference.Hochberg correction
Hommel correction