Background
Lord’s paradox presents a fascinating challenge in causal inference and statistics. It highlights how different statistical methods applied to the same data can lead to contradictory conclusions. The paradox typically arises when comparing changes over time between two groups in the context of adjusting for baseline characteristics. Despite its theoretical nature, the phenomena is deeply relevant to practical data analysis, especially in observational studies where causation is challenging to establish. Let’s look at an example.
Diving Deeper
Mean Differences Over Time
To explore Lord’s paradox, consider the following scenario: Suppose we have two groups of individuals— and —with their body weights measured at two time points: before and after a specific intervention. Let the weight at the initial time point be denoted as , and the weight at the final time point be . We are interested in whether the intervention caused a change in weight between the two groups.
One approach to analyze this is to compute the (unadjusted) mean weight change for each group over time:
If this quantity is statistically significant, we might conclude that the intervention had a differential effect.
Controlling for Baseline Characteristics
An alternative approach involves adjusting for baseline weight using, for example, a regression model:
where is a binary indicator for group membership and is an error term. Here, captures the group difference in , linearly controlling for baseline body weight.
Surprisingly, these two approaches can yield conflicting results. For example, the former method might suggest no difference, while the regression adjustment indicates a significant group effect.
An Explanation
This contradiction arises because the two methods implicitly address different causal questions.
- Method 1 asks: “Do Groups and gain/lose different amounts of weight?”
- Method 2 asks: “Given the same initial weight, does any of the groups end up at different final weights?”
The regression approach adjusts for baseline differences, assuming is a confounder.
The key insight is that the choice of analysis reflects underlying assumptions about the causal structure of the data. If is affected by group membership (e.g., due to selection bias), then adjusting for it may introduce bias rather than remove it. However, in practice we most often should control for baseline characteristics as they are critical in balancing the treatment and control groups.
The Simpson’s Paradox Once Again
I recently illustrated the more commonly discussed Simpson’s paradox. Interestingly, a 2008 paper claims that two phenomena are closely related, with the Lord’s paradox being a “continuous version” of Simpson’s paradox.
An Example
Let’s look at some code illustrating Lord’s paradox in R
. We start with simulating a dataset where two groups have identical distributions of and , yet differing relationships between the two variables.
rm(list=ls())
set.seed(1988)
n <- 1000
# Simulate data for two groups (e.g., male/female students)
group <- factor(rep(c("A", "B"), each = n/2))
# Initial weight (pre). # Group A starts with higher average weight
weight_pre <- numeric(n)
weight_pre[group == "A"] <- rnorm(n/2, mean = 75, sd = 10)
weight_pre[group == "B"] <- rnorm(n/2, mean = 65, sd = 10)
# Final weight (post). Both groups improve by the same amount on average
gain <- rnorm(n, mean = 10, sd = 5)
weight_post <- weight_pre + gain
# Create data frame
data <- data.frame(group = group, pre = weight_pre, post = weight_post)
# Analysis 1: Compare change scores between groups
t.test(post - pre ~ group, data = data)
> p-value = 0.6107
# Analysis 2: Regression Adjustment
model <- lm(post ~ group + pre, data = data)
summary(model)
> p-value = 0.08428742
The former approach shows lack of statistically significant differences in the body weight after the intervention between the two groups (p-value = ). The results from the latter method do show meaningful differences (p-value = ).
This illustrates the core of Lord’s paradox – the statistical approach chosen can lead to different interpretations of the same underlying phenomenon.
You can find this code in this GitHub repo.
Takeaways
- Lord’s paradox underscores the importance of aligning statistical methods with causal assumptions.
- Different methods answer different questions and may yield contradictory results if applied blindly.
- Careful consideration of the data-generating process and the role of potential confounders is crucial in choosing the appropriate analytical approach.
References
Lord, E. M. (1967). A paradox in the interpretation of group comparisons. Psychological Bulletin, 68, 304–305. doi:10.1037/h0025105
Lord, F. M. (1969). Statistical adjustments when comparing preexisting groups. Psychological Bulletin, 72, 336–337. doi:10.1037/h0028108
Lord, E. M. (1975). Lord’s paradox. In S. B. Anderson, S. Ball, R. T. Murphy, & Associates, Encyclopedia of Educational Evaluation (pp. 232–236). San Francisco, CA: Jossey-Bass.
Tu, Y. K., Gunnell, D., & Gilthorpe, M. S. (2008). Simpson’s Paradox, Lord’s Paradox, and Suppression Effects are the same phenomenon–the reversal paradox. Emerging themes in epidemiology, 5, 1-9.
Leave a Reply