In data analysis, it's common to run many hypothesis tests at once for example, checking thousands of genes in a medical study to see which are linked to a disease. But the more tests we do, the higher the chance of getting results that look important just by accident. To reduce these false positives, statisticians use correction methods. The Holm-Bonferroni method is one of them and is popular because it controls these errors without being too strict, making it more effective than the older Bonferroni method.
How Does It Work?
Let’s say you run m hypothesis tests and get m p-values.
1. Sort the p-values in ascending order:
p_{(1)} \leq p_{(2)} \leq \cdots \leq p_{(m)}
2. Assign ranks to the sorted p-values. Let p_{(i)} be the i-th smallest p-value.
3. Compare each p-value to its adjusted threshold:
p_{(i)} \leq \frac{\alpha}{m - i + 1}
where α\alphaα is your desired significance level (e.g., 0.05).
4. Reject all hypotheses from the first p_{(1)} up to the last one that satisfies the condition.
Once a p-value fails the condition, stop rejecting further hypotheses, even if the later p-values meet the threshold.
Example- Use the Holm-Bonferroni method to test the following four hypotheses and their associated p-values at an alpha level of 0.05:
Hypothesis | p-value |
---|
H1 | 0.01 |
---|
H2 | 0.04 |
---|
H3 | 0.03 |
---|
H4 | 0.005 |
---|
Step-by-Step Solution Using the Holm-Bonferroni Method
Step 1: Sort the p-values in ascending order and match the corresponding hypotheses
Hypothesis | p-value |
---|
H4 | 0.005 |
---|
H1 | 0.01 |
---|
H3 | 0.03 |
---|
H2 | 0.04 |
---|
Step 2: Compute adjusted significance levels
Hypothesis | p-value | \alpha |
---|
H4 | 0.005 | 0.0125 |
---|
H1 | 0.01 | 0.0167 |
---|
H3 | 0.03 | 0.025 |
---|
H2 | 0.04 | 0.05 |
---|
Step 3: Compare p-values to corresponding α,
Hypothesis | p-value | \alpha |
---|
H4 | 0.005 | Reject |
---|
H1 | 0.01 | Reject |
---|
H3 | 0.03 | Do not reject |
---|
H2 | 0.04 | Do not reject |
---|
Holm-Bonferroni vs Bonferroni
Method | Approach | Conservative? | Power |
---|
Bonferroni | Fixed threshold | Very high | Low |
---|
Holm-Bonferroni | Stepwise, adaptive | Less conservative | Higher |
---|
- Bonferroni rejects only if p-value < α / m
- Holm-Bonferroni adapts the threshold as you go through each p-value
So, Holm-Bonferroni is always at least as powerful as Bonferroni, and usually better.
When to Use Holm-Bonferroni
- When controlling the family-wise error rate (FWER) is important
- When you have a moderate number of hypothesis tests
- When you want a method that's stronger than Bonferroni but still simple to understand and apply
Advantages of Holm-Bonferroni
- More powerful than Bonferroni (less likely to miss true positives)
- Easy to implement by sorting and comparing p-values
- Controls FWER in a reliable way
- Works for independent and some dependent tests
Limitations
- Still conservative, especially when many tests are performed
- Does not control the False Discovery Rate (FDR) use Benjamini-Hochberg for that
- Cannot recover power loss in very high-dimensional testing
Similar Reads
Bonferroni Test In statistical analysis, multiple hypothesis testing is a frequent challenge, particularly in experiments or studies with numerous variables. One common issue is the increased likelihood of committing a Type I error (false positive) as the number of tests grows. To address this, the Bonferroni Test
4 min read
How to Perform a Bonferroni Correction in R? The Bonferroni correction is a method to adjust p-values when performing multiple comparisons to control the family-wise error rate. It is particularly useful in hypothesis testing where multiple tests are conducted simultaneously. Here's how you can perform a Bonferroni correction in R Programming
3 min read
How to Perform a Bonferroni Correction in Excel? The Bonferroni correction, also known as the Bonferroni type adjustment, is one of the most fundamental processes used in multiple comparison testing. The name was inspired by Carlo Emilio, the museum's Italian curator. The procedure of altering the alpha (α) level for a series of statistical tests
4 min read
Jacobian Method The Jacobian Method, also known as the Jacobi Iterative Method, is a fundamental algorithm used to solve systems of linear equations. This method, named after the mathematician Carl Gustav Jacob Jacobi, is particularly useful when dealing with large systems where direct methods are computationally e
7 min read
Newton's method in Machine Learning Optimization algorithms are essential tools across various fields, ranging from engineering and computer science to economics and physics. Among these algorithms, Newton's method holds a significant place due to its efficiency and effectiveness in finding the roots of equations and optimizing functi
15 min read
What is mean method formula? Statistics is a mathematical branch that is carried out by the collection and summarization of data. It is concerned with collecting, analyzing, interpreting, presenting a set of data. Statistics has its role in the field of data collection especially data used by the government like a population ce
3 min read