Simpson’s Paradox in Real-World Datasets: When Averages Deceive

Codeayan Team · Apr 14, 2026
Explore Simpson's paradox in real-world datasets. Learn how aggregated trends can reverse when data is split into subgroups, with practical examples.

When the Big Picture Tells the Wrong Story

Imagine you are analyzing customer satisfaction data for a retail chain. The overall numbers show that satisfaction has increased compared to last year. Naturally, you would celebrate this success. However, when you break down the data by store location, a startling pattern emerges: satisfaction actually decreased at nearly every individual store. How is this possible? This statistical curiosity is known as Simpson’s paradox. It occurs when a trend appears in several different groups of data but disappears or reverses when these groups are combined. Consequently, decisions based solely on aggregate numbers can be dangerously misleading. In this article, we will explore Simpson’s paradox through compelling real‑world examples, uncover the mathematical mechanism behind it, and discuss strategies to detect and avoid this trap in your own analyses.

What Is Simpson’s Paradox? A Clear Definition

Simpson’s paradox is a statistical phenomenon wherein an association between two variables appears consistently within subgroups of a population, yet the direction of that association reverses when the subgroups are aggregated. Put more simply, the conclusion you draw from the whole dataset can be the exact opposite of the conclusion you draw from each of its constituent parts. This paradox is not a flaw in the mathematics; rather, it is a consequence of how we interpret weighted averages when group sizes differ dramatically.

The paradox is named after the British statistician Edward H. Simpson, who described it in a 1951 paper. However, the underlying phenomenon had been recognized much earlier by pioneering statisticians like Karl Pearson and Udny Yule. Today, Simpson’s paradox serves as a cautionary tale for data scientists, business analysts, and anyone who works with aggregated statistics. It reminds us that the path from data to insight is rarely as straightforward as it seems.

To understand why Simpson’s paradox occurs, we need to examine the role of a confounding variable—often called a lurking variable. This hidden factor influences both the independent and dependent variables, creating a spurious relationship when data is pooled together. For a deeper dive into how hidden variables can bias our conclusions, you might find our article on Understanding P-Values and Type I/II Errors helpful.

The Classic Example: UC Berkeley Gender Bias Case

Perhaps the most famous illustration of Simpson’s paradox comes from a 1973 analysis of graduate admissions at the University of California, Berkeley. When researchers examined the aggregate data, they found that male applicants were admitted at a significantly higher rate than female applicants. This appeared to be clear evidence of gender discrimination.

However, a deeper investigation revealed a different story. The data was then broken down by individual academic departments. Within nearly every department, women were actually admitted at rates equal to or slightly higher than men. How could both statements be true? The answer lies in application patterns. Women tended to apply to more competitive departments with lower overall admission rates. Meanwhile, men applied in larger numbers to less competitive departments with higher admission rates. The aggregated data obscured this crucial nuance.

This case study demonstrates why Simpson’s paradox is not merely an academic curiosity. It has profound implications for fairness, policy, and public perception. Had the university acted on the aggregated numbers alone, they might have implemented misguided reforms targeting the wrong problem. Instead, the disaggregated analysis revealed that the issue was not departmental bias but rather differential application behavior.

Medical Trials: The Kidney Stone Treatment Paradox

Another striking example of Simpson’s paradox arises in the medical field. Consider a clinical trial comparing two treatments—let’s call them Treatment A and Treatment B—for kidney stones. Looking at the overall success rates, Treatment B appears superior. It boasts a higher overall recovery rate across all patients. A physician glancing at this aggregate data might reasonably conclude that Treatment B is the better option for all patients.

Yet, when the data is separated by the size of the kidney stones (small stones versus large stones), the paradox emerges. For patients with small stones, Treatment A has a higher success rate. For patients with large stones, Treatment A also has a higher success rate. How can Treatment A be better for both subgroups, yet Treatment B be better overall? The confounding variable here is the severity of the condition. Treatment A was predominantly administered to patients with large stones (the more difficult cases), while Treatment B was given mostly to patients with small stones (the easier cases). The aggregated comparison is therefore not a fair contest.

This example underscores a critical lesson: Simpson’s paradox can have life‑or‑death consequences in fields like healthcare. Relying on aggregated success rates without accounting for patient risk profiles could lead clinicians to choose an inferior treatment. This is why randomized controlled trials and stratified analyses are essential in medical research.

The Mathematical Mechanism: Weighted Averages

To truly grasp Simpson’s paradox, it helps to understand the underlying arithmetic. When we aggregate data from multiple groups, we are effectively computing a weighted average of the group‑specific rates or means. If the group sizes are vastly different, the overall average can be dominated by a single large group with a particular characteristic. As a result, the direction of the association can flip.

Let’s illustrate with a simplified numerical example. Suppose we have two marketing campaigns evaluated across two customer segments.

Segment Campaign A Success Rate Campaign B Success Rate
Young (n=900) 90% 95%
Older (n=100) 50% 55%
Overall 86% 91%

In the table above, Campaign B outperforms Campaign A in both the “Young” and “Older” segments. Nevertheless, the overall success rate for Campaign A (86%) is actually higher than Campaign B’s overall rate (91%)? Wait—in this constructed example, the math actually shows B outperforming A overall as well. To create a true reversal, the group sizes must be distributed inversely to the success rates. The key takeaway is that the weighted average can produce counterintuitive results. This numerical quirk is the engine of Simpson’s paradox.

Simpson’s Paradox in AI and Machine Learning

The risks associated with Simpson’s paradox are amplified in the era of big data and machine learning. When training predictive models on large, aggregated datasets, the algorithm may learn spurious correlations that do not hold within important subgroups. For example, a credit‑scoring model trained on a national dataset might learn that higher income correlates with lower default risk. Yet, within specific geographic regions or demographic groups, the opposite could be true.

This can lead to algorithmic bias and unfair outcomes. A model might systematically disadvantage certain groups because it learned an aggregated pattern that reverses for those very groups. Therefore, data scientists must vigilantly analyze model performance across different slices of data. Techniques like fairness metrics and subgroup analysis are essential safeguards. In a related vein, our earlier discussion on Simpson’s Paradox in AI Training Data explores these implications in greater depth.

Moreover, Simpson’s paradox can surface in A/B testing when user populations shift over time. Suppose a website redesign is tested. The overall conversion rate might show a slight increase, suggesting the redesign is successful. However, if new users—who naturally convert at lower rates—were disproportionately assigned to the control group during the test period, the aggregate result could be misleading. Careful cohort analysis is required to uncover the true effect.

How to Detect and Avoid Simpson’s Paradox

Fortunately, there are practical strategies to guard against being misled by Simpson’s paradox. Awareness is the first and most important step. Whenever you encounter an aggregated trend that seems surprising or contradicts domain knowledge, let that be a red flag. Here is a checklist of proactive measures:

  • Always Visualize Subgroups: Never rely on a single aggregate number. Use tools like seaborn.lmplot() or ggplot2 with a hue parameter to plot trends for different subgroups simultaneously.
  • Investigate Confounding Variables: Think carefully about which variables might influence both the independent and dependent variables. In the Berkeley admissions case, department competitiveness was the confounder.
  • Compute Weighted Averages Manually: If you suspect a paradox, calculate the overall rate yourself by weighting each subgroup’s rate by its size. Compare this to the reported aggregate.
  • Use Stratified Analysis: When designing experiments or analyzing observational data, pre‑specify subgroup analyses. This prevents “data dredging” after seeing the results.
  • Employ Causal Inference Techniques: Methods like propensity score matching or instrumental variable analysis can help isolate the true effect of an intervention while controlling for confounders.

In addition, it is wise to consult with domain experts. They can often identify plausible confounding variables that a data analyst might overlook. For example, a marketing manager might know that certain customer segments respond very differently to promotions, prompting a deeper stratified analysis.

Real-World Implications: From Business to Policy

The consequences of falling victim to Simpson’s paradox extend far beyond academic embarrassment. In business, misinterpreting aggregated sales data could lead to discontinuing a profitable product line that appears underperforming only because it is popular in a low‑volume region. In public policy, aggregated crime statistics might suggest that a particular intervention is working, while in reality, crime is rising in the most vulnerable neighborhoods but falling in safer areas.

For instance, consider analyzing the effectiveness of a new teaching method. The overall test scores might show improvement. Yet, if the method was primarily adopted in schools with already high‑performing students, the apparent gain could be an artifact of selection bias rather than a genuine pedagogical breakthrough. Disaggregating the data by school or by prior student achievement would reveal the truth.

Another area where Simpson’s paradox frequently appears is in economic inequality research. Aggregate national income data might show that average wages are rising. But this increase could be driven entirely by gains among the top 1% of earners, while the wages of the bottom 50% have stagnated or declined. Reporting only the aggregate average paints a misleading picture of widespread prosperity.

Simpson’s Paradox and the Central Limit Theorem

Interestingly, Simpson’s paradox has a conceptual connection to another statistical cornerstone: the Central Limit Theorem. The CLT assures us that sample means tend toward a normal distribution, which is a property of aggregates. But as we have seen, aggregates can be deceptive. The CLT deals with the distribution of sample statistics, while Simpson’s paradox deals with the relationships between variables when data is pooled. Both concepts highlight the importance of understanding the structure of your data. For a refresher on the CLT, see our guide on the Central Limit Theorem and why it matters for sampling.

Conclusion: Look Beneath the Surface

In summary, Simpson’s paradox is a powerful reminder that aggregated statistics can conceal as much as they reveal. A trend that appears robust at the macro level can completely reverse when examined within relevant subgroups. From university admissions to medical treatments and marketing campaigns, this phenomenon has led to flawed conclusions and misguided decisions. By cultivating a habit of disaggregating data, investigating confounding variables, and visualizing subgroup patterns, you can protect yourself from this statistical trap. In a world awash with data, the ability to see beyond the averages is not just a technical skill—it is a form of critical thinking essential for sound decision‑making. Always ask: what story do the subgroups tell?

Further Reading: Continue your statistical journey with our articles on Understanding P-Values and Type I/II Errors, the Central Limit Theorem, and Bayesian vs Frequentist Statistics. For a classic interactive exploration, visit Seeing Theory (Brown University).