Educational Attainment Polygenic Scores Track Civilization Stage, Not Just Chronology
A quantitative analysis of ancient DNA and archaeological stage
Ancient DNA contains a surprise. When researchers extract polygenic scores linked to educational attainment from prehistoric skeletons, the scores rise through time. Hunter-gatherers score lower than early farmers, early farmers lower than Bronze Age populations, and so on. The standard interpretation is that this reflects something about the passage of time itself such as selection, migration, demographic turnover, or some mixture of all three.
But time is not the only thing changing across that span. Social and technological complexity are changing too. A Paleolithic forager band and an Iron Age kingdom are not just separated by millennia. They are separated by farming, writing, cities, specialization, storage, hierarchy, and the entire accumulated weight of what we loosely call civilization. So here is the question worth asking: when those polygenic scores rise through ancient history, are we really just watching a clock tick or are we watching something track the emergence of more complex ways of organizing human life?
Using ancient individuals from the AADR dataset, I assigned each archaeological period a civilization-stage score and asked whether that score predicts educational-attainment polygenic scores even after controlling for absolute date. In practice, that meant coding Paleo-Mesolithic groups as 1, Neolithic groups as 2, Copper Age groups as 3, Bronze Age groups as 4, and Iron Age groups as 5.
The leverage comes from a basic fact of world prehistory: civilization and chronology do not move in lockstep. Greece had farming millennia before Britain. Some populations at the same date occupied very different social worlds. That mismatch is what makes it possible to pull the two apart and ask which one is actually doing the work.
A time trend by itself could reflect selection, migration, demography, drift, or some mixture of all of them. The harder question is whether the relevant axis is not just time, but civilization.
If you take two ancient populations from different social worlds, but do not let chronology do all the explanatory work, does their place on the civilizational ladder still matter?
The modern world has made one thing hard to miss. Cognitive performance, schooling, institutional complexity, and economic development cluster together. But once we move back into deep history, the picture becomes much hazier. We still know remarkably little about whether the genetic variants associated with educational attainment track not just the passage of time, but the emergence of more complex forms of social organization.
Gregory Clark (2007) argued that preindustrial societies may have undergone differential reproduction in ways that slowly shifted the distribution of traits favorable to economic success. Galor and Moav (2002) built a related idea into a formal evolutionary growth framework. In their model, the Malthusian world did not simply hold humanity down. It also created selection pressures that gradually favored traits complementary to human capital, technology, and later economic takeoff. On that view, the Industrial Revolution was not a bolt from the blue. It was the late visible phase of a much longer process.
I'm not testing their mechanism directly but the question points in the same direction. If traits linked to educational attainment were historically relevant to the capacity of populations to sustain more complex forms of life, then archaeological stage should retain some predictive power even after absolute time is held constant. In that sense, the civilization-stage score is crude, but it is also useful. It is a rough proxy for where a population stood in the long transition from simple subsistence regimes to more knowledge-intensive and organizationally demanding societies.
The Neolithic did not begin everywhere at the same moment. Greece had early farming millennia before the British Isles. Northern China followed a different timetable again. The same is true of later transitions. At any given date, some populations were still living in relatively simple social worlds while others had already crossed into much more complex ones. And populations grouped into the same broad archaeological stage could be separated by very large spans of time.
A map of the spread of agriculture makes the point immediately. The transition to farming did not unfold at a uniform pace across Europe. Some regions crossed into the Neolithic much earlier than others, while large parts of the continent remained outside it for centuries or millennia longer. That staggered transition is exactly what makes the present analysis possible. If different populations occupied different civilizational stages at the same broad date, then archaeological stage and chronology can, at least in principle, be pulled apart.

The sample is global in scope, not limited to Europe, which provides the statistical power needed to detect meaningful mismatches.
That mismatch is what makes the whole exercise possible. Once we exploit the fact that date and stage are misaligned, we can ask the real question: when educational-attainment polygenic scores rise through ancient time, are we just looking at chronology, or are we also looking at civilization?
How I tested it
The data come from ancient individuals in the public AADR 1240k dataset — the largest collection of ancient human genomes currently available. After standard cleaning, I was left with samples spanning roughly 45,000 years of prehistory across Europe, the Near East, and Central Asia.
I first collapsed the archaeological labels into five broad stages: Paleo-Mesolithic, Neolithic, Copper Age, Bronze Age, and Iron Age. I then coded them numerically from 1 to 5. Paleo-Mesolithic foragers were coded as 1, Neolithic farmers as 2, Copper Age societies as 3, Bronze Age societies as 4, and Iron Age societies as 5. The logic was straightforward: each step up the scale marks a broad increase in sedentism, productive technology, storage, specialization, hierarchy, and the capacity to accumulate and transmit knowledge. Historical and medieval cases were not used as a separate final stage, because that category is too heterogeneous and not archaeologically comparable in the same clean way.
The 1 to 5 scale should not be read as a literal ruler of civilization. It is a rough ordinal ladder. Its only strong claim is that these stages can be placed in a meaningful sequence from simpler to more complex forms of social organization. The regression then asks whether moving upward along that ladder is associated with higher polygenic scores on average.
The outcome variable was each individual’s educational-attainment polygenic score, a measure derived from their ancient DNA that aggregates hundreds of small genetic variants associated with years of schooling in modern populations.
The core question was simple: does civilization stage predict that score even after I control for when the individual lived? I ran the analysis across two time windows — the full 45,000-year span, and the Holocene alone (the last 12,000 years) — to see whether the result depended on the deepest contrast between foragers and everyone else. All models controlled for date, genomic data quality, and geography. The main specifications also controlled for ancestry, so that broad population movements couldn’t masquerade as a civilization effect.
Time and civilization are related, but not the same thing
Archaeological stage and calendar date are correlated, but far from perfectly. Populations living at similar dates can occupy very different civilizational positions, and populations assigned to the same broad stage can be separated by long stretches of time. This can be seen in the figure below.
Mismatch Between Time and Civilization Stage
If civilization stage predicts EA-linked polygenic scores after time is held constant, then the result cannot be reduced to a simple temporal climb.


