◆ Powerful

The Ecological Fallacy

Using statistics about groups to draw conclusions about individuals is a structural error with serious real-world consequences — in criminal profiling, public health policy, and cross-national research. Here is how to recognise it and its mirror image.

Time: 12 minutes
Requires: Unit 3.7

Opening Hook

In 2012, a paper appeared in the New England Journal of Medicine. The author, Dr Franz Messerli of Columbia University, had plotted the per-capita chocolate consumption of twenty-three countries against the number of Nobel Prize winners each country had produced per ten million people. The correlation was striking: r = 0.791. The more chocolate a country ate, the more Nobel laureates it produced. The paper was titled “Chocolate Consumption, Cognitive Function, and Nobel Laureates.”

The New England Journal of Medicine is the most prestigious medical journal in the world. The study was covered by the New York Times, the BBC, and dozens of other outlets. Some readers got the joke. Many did not. The paper was received as genuine evidence that eating chocolate improves brain function.

Here is what the paper actually showed. Countries with high chocolate consumption also tend to be wealthy, northern European countries. Wealthy, northern European countries also tend to produce a lot of Nobel laureates. The country-level data tells you something about countries. It tells you nothing about what happens inside the brain of any individual chocolate eater. The person consuming the relevant data — a reader of the New England Journal of Medicine — is an individual, not a country. But the data being presented is about countries. The gap between those two levels of analysis is where the deception lives.

This is the ecological fallacy.

The Concept

The ecological fallacy is the error of using statistics calculated at the group level to draw conclusions about individuals. The word “ecological” here does not refer to the natural environment. It refers to populations, or ecological units, as distinct from the individuals within them. The error is inferring something about one level of analysis (the person) from data collected at a different level of analysis (the neighbourhood, the country, the ward, the region).

The foundational demonstration came in 1950, when the American sociologist W.S. Robinson published a paper in the American Sociological Review. Robinson took US census data and looked at the relationship between race and illiteracy. At the level of US states, the correlation between the proportion of Black residents and the illiteracy rate was very high: around 0.95. If you had looked only at that number, you might conclude that race and illiteracy were tightly linked. But when Robinson computed the same correlation at the individual level — looking at actual people rather than states — the correlation dropped to around 0.20. The state-level finding and the individual-level finding were pointing to something quite different. The state-level correlation was driven by historical geography: states with higher proportions of Black residents were, in 1930, largely southern states with systematically underfunded education for all their residents. The high ecological correlation was picking up the legacy of policy, not a characteristic of individuals. Using the state-level number to say something about any particular person would have been not just wrong, but badly wrong in a way that had real consequences.

A closely related phenomenon is Simpson’s Paradox, named after the statistician E.H. Simpson who described it formally in 1951. Simpson’s Paradox occurs when a trend that appears in all subgroups of a data set disappears, or reverses, when the groups are combined. The Berkeley admissions case is the most widely cited example. In the autumn of 1973, Berkeley’s graduate school admitted around 44 percent of male applicants and 35 percent of female applicants. Looked at in aggregate, this appeared to show systematic bias against women. When the statistician Peter Bickel disaggregated the data by department, a different picture emerged: in four out of six departments, women were actually more likely to be admitted than men. What was happening was that women had applied in larger numbers to highly competitive departments (like English) that admitted a small fraction of all applicants regardless of gender, while men had applied in larger numbers to departments with higher admission rates (like mechanical engineering). The aggregate number was technically accurate and completely misleading. The paradox arises whenever subgroups differ not just in the outcome of interest but in some underlying variable that itself affects the outcome.

There is also a mirror-image error called the atomistic fallacy. Where the ecological fallacy goes from group to individual, the atomistic fallacy goes from individual to group. If you know that a particular person who exercises daily is in excellent health, you cannot conclude that a neighbourhood where most people exercise will therefore have excellent aggregate health outcomes. Individual health data does not simply aggregate into group health. Social determinants, infrastructure, collective stress, and resource availability operate at the group level and are not reducible to the sum of individual characteristics. Both errors have the same root structure: they assume that patterns found at one level of analysis will transfer cleanly to another level. They almost never do.

Cross-national comparisons are a rich source of ecological fallacies precisely because the data is readily available and the temptation to conclude something about individuals from national-level patterns is very strong. Countries with higher internet access tend to have lower rates of some diseases. Countries with higher GDP per capita tend to have lower rates of many adverse outcomes. Countries with different dietary patterns have different rates of certain cancers. Every one of these observations may be true at the national level and tell you almost nothing about whether any individual in any of those countries should change their behaviour. The national-level correlation could be driven entirely by infrastructure, inequality, climate, healthcare access, or any number of other structural factors. To take the national average and apply it to an individual is to confuse the map for the territory.

Why It Matters

The ecological fallacy is not an abstract statistical point. It shows up in three areas of public life with significant consequences.

The first is area-based health policy. Public health services in the UK and elsewhere routinely use deprivation indices, which assign scores to small geographic areas (electoral wards, census tracts, postcodes) based on factors like unemployment, income, housing quality, and education levels. These indices are useful for allocating resources at scale. The fallacy enters when the area score is applied to individuals in that area. A 2002 study in the Journal of the Royal Statistical Society, by Lancaster and Green, examined exactly this problem. They compared individual-level health data with ward-level deprivation scores from the 1991 UK census and found substantial divergence. Many individuals living in deprived wards are not themselves deprived. Many individuals living in affluent wards face significant hardship. Using the area score as a proxy for the individual’s situation produces both false positives (treating non-deprived individuals as deprived) and false negatives (missing deprived individuals in relatively affluent areas). More recent research, published in the Journal of Public Health in 2025, confirmed that household income predicts child health outcomes significantly better than neighbourhood deprivation scores. The individual measure outperforms the ecological one. This matters because it shapes which interventions get funded, which populations get screened, and which people get missed.

The second area is criminal profiling and law enforcement. When an area has a higher crime rate, this is a statement about events in a geography, not a statement about the inhabitants of that geography. Using area-level crime statistics to treat every resident as a suspect is the ecological fallacy operationalised as policy. An individual who lives in a high-crime area may have no elevated probability of committing crime. The correlation between area crime rates and individual behaviour is always much weaker than the area statistics suggest, for exactly the reasons Robinson identified in 1950: the area statistic reflects structural conditions (poverty, unemployment, housing quality, policing patterns, historical disinvestment) more than it reflects individual propensity.

The third area is cross-national policy borrowing. Governments frequently look at other countries and conclude that policies which are associated with good outcomes over there should be imported. Finland’s educational system produces high PISA scores. South Korea has low rates of a particular disease. Denmark scores highly on happiness indices. Each of these is a statement about a national aggregate. The structural, cultural, historical, and institutional differences between countries are so substantial that the ecological correlation between a policy and an outcome in one country provides only weak evidence about what would happen if you implemented the same policy in a different country. This does not mean cross-national comparisons are worthless. It means they are starting points for inquiry, not conclusions.

How to Spot It

The tell for the ecological fallacy is a gap between the level at which data was collected and the level at which a claim is being made.

The documented case that set the standard is Robinson (1950). Robinson himself spelled out the tell: he showed that you can have a very high ecological correlation and a very low individual-level correlation for the same pair of variables, in the same dataset, at the same moment in time. The state-level number (0.95) was not a weak version of the individual-level number (0.20). It was measuring something categorically different. The fact that two numbers are called by the same name (correlation between race and literacy) does not make them the same thing. One is about geography, the other is about people.

When you encounter a statistical claim, the question to ask is: at what level was this data collected? If the data is about countries, regions, cities, areas, wards, or any other geographic or institutional unit, and the conclusion is about individuals within those units, the ecological fallacy may be operating. Ask whether the relationship found between group averages is likely to hold within any particular group, or whether it is driven by between-group differences that say nothing about individual variation.

The secondary tell is a very high correlation. Ecological correlations tend to be higher than individual-level correlations because averaging smooths out individual variation. When you see a strikingly clean correlation in data aggregated across units, this is a reason for more scepticism, not less. The Messerli chocolate paper had a correlation of 0.79. That number should raise an eyebrow, not lower a guard. Very high correlations in cross-national or cross-regional data are usually a sign that both variables are being driven by some underlying structural factor, not that one causes the other at the individual level.

Your Challenge

A study published in a public health journal examines US states. It finds that states with higher rates of gun ownership also have higher murder rates. The correlation is statistically significant and has been replicated in multiple studies. A commentator argues that this proves individual gun ownership causes homicide. A gun rights advocate responds that the vast majority of individual gun owners never commit any crime, so the statistic cannot tell us anything about individuals.

One of these people is making the ecological fallacy. The other is making the atomistic fallacy. Which is which? What would you need to know to move from the state-level correlation to a well-founded claim about what happens at the individual level? And what is the third variable problem lurking behind both positions?

There is no answer on this page. That is the point.

References

Messerli, F.H. “Chocolate Consumption, Cognitive Function, and Nobel Laureates.” New England Journal of Medicine 367 (2012): 1562–1564. https://www.nejm.org/doi/abs/10.1056/NEJMon1211064

Robinson, W.S. “Ecological correlations and the behavior of individuals.” American Sociological Review 15, no. 3 (1950): 351–357. Discussed and analysed in: Freedman, D.A., “Ecological Inference and the Ecological Fallacy,” Stanford University. https://web.stanford.edu/class/ed260/freedman549.pdf

Bickel, P.J., Hammel, E.A., and O’Connell, J.W. “Sex Bias in Graduate Admissions: Data from Berkeley.” Science 187, no. 4175 (1975): 398–404. Summarised at refsmmat.com: https://www.refsmmat.com/posts/2016-05-08-simpsons-paradox-berkeley.html

Lancaster, G.A. and Green, M. “Deprivation, ill-health and the ecological fallacy.” Journal of the Royal Statistical Society: Series A (Statistics in Society) 165, no. 2 (2002): 263–278. https://academic.oup.com/jrsssa/article/165/2/263/7084157

Fieldhouse, E.A. and Tye, R. “Deprived People or Deprived Places? Exploring the Ecological Fallacy in Studies of Deprivation with the Samples of Anonymised Records.” Environment and Planning A 28, no. 2 (1996): 237–259. https://journals.sagepub.com/doi/abs/10.1068/a280237

Oxford Academic Journal of Public Health, “Does household income predict health and educational outcomes in childhood better than neighbourhood deprivation?” (2025). https://academic.oup.com/jpubhealth/article/47/1/62/7903429