Counting the Uncountable - The Limits of Quantification in the Social Sciences: Why the human world resists the logic of numbers

The debate over appropriate methods in the social and human sciences is as old as the disciplines themselves. Since the early 20th century, when psychology and later the social sciences began adopting methods from the natural sciences, a question has persisted: Can social phenomena truly be captured with numbers? Or does the quantitative approach hit an insurmountable boundary here? In this article, I want to explore three key problem areas that demonstrate why quantification in the social sciences is fundamentally problematic: construct validity, the generalizability crisis, and the limitations of RCTs (Randomized Controlled Trials).

1. Construct Validity: Are We Measuring the Right Thing?

One of the earliest and most pointed insights into this problem came from Lee J. Cronbach, who introduced the concept of construct validity in 1955. Ironically, this is the same Cronbach who developed “Cronbach’s Alpha,” a standard measure of reliability. He emphasized that the crucial question is not how reliably a test measures, but whether it measures what it claims to measure at all.

In both psychology and the social sciences, many research objects are abstract constructs: “intelligence,” “social trust,” “democratic satisfaction,” or “prejudice.” Such terms have no direct material counterpart. They cannot be measured like temperature or weight. Instead, indicators are constructed—such as questions in a survey or reaction times in an experiment. Yet whether these indicators truly capture the construct is highly uncertain.

Example: Intelligence tests measure specific skills like pattern recognition or verbal comprehension. But whether they truly capture “intelligence” as a whole is questionable. Similarly, in sociology, “social capital” may be operationalized via membership in clubs or associations. Reducing complex constructs to quantifiable aspects obscures their actual complexity.

Cronbach called this the “needle’s eye of construct validity”: all empirical research in the human sciences must pass through this narrow point, and many approaches fail right here.


2. The Generalizability Crisis: When Results Depend on Context

Even if we assume a construct is correctly measured, the next hurdle remains: generalizability. A study may produce valid results in a specific context—but do these results hold beyond that context? Here, the social sciences face a systematic challenge.

In the natural sciences, universal laws are often possible because the objects studied (e.g., chemical reactions) behave consistently regardless of context. In the social sciences, however, people are embedded in social, cultural, and historical contexts that cannot be fully controlled.

This gives rise to what is called the generalizability crisis: results cannot easily be applied to other groups, cultures, or time periods. A psychological experiment conducted at an American university tells us very little about people in other countries or social milieus. Likewise, a sociological survey from the 1980s may yield entirely different results today due to changing societal conditions.

A classic example is the many social psychology studies based on WEIRD populations (Western, Educated, Industrialized, Rich, Democratic). These studies are often presented as universal but, in reality, only describe a very specific, culturally influenced segment of humanity. The social sciences’ claim to generate universal insights is fundamentally undermined here.


3. RCTs in the Social Sciences: Illusory Precision through Experiments

Randomized Controlled Trials (RCTs) are considered the “gold standard” of empirical research. In medicine, they are invaluable because clearly defined interventions and control groups are possible. For several decades, RCTs have increasingly been applied in the social sciences, for example in development economics or education. Yet here the methodological limitations become especially evident.

The basic principle of RCTs sounds deceptively simple: participants are randomly assigned to an intervention or control group, and differences are measured. This is supposed to allow causal relationships to be clearly identified. But in the social sciences, this approach only works to a limited extent:

  1. Control is impossible: Humans cannot be isolated like chemical substances in a lab. Their behavior is influenced by countless contextual factors that an RCT cannot capture.
  2. External validity is weak: Even if an RCT shows that a particular intervention works in a rural school in Kenya, this does not mean it will work in a German city.
  3. Ethical boundaries: Many of the most pressing social questions—poverty, violence, trauma—cannot be randomly assigned ethically.

Economist Esther Duflo and her team gained worldwide attention with RCTs aimed at poverty reduction. Yet critics such as Angus Deaton and Nancy Cartwright have pointed out that these studies often deliver only “illusory precision”: they show small, context-specific effects that cannot be generalized. RCTs suggest precision where, in reality, context-dependence dominates.

Picture: thanks to Alex Shute on Unsplash 

Conclusion: Thinking Beyond Empiricism

The three examples—construct validity, the generalizability crisis, and RCTs—highlight a common underlying problem: the social cannot be fully quantified. Of course, we can count people, sum survey responses, or measure reaction times. But social phenomena themselves—trust, values, motives, culture—are not material objects that can be neatly represented by numbers.

For the past century, the social sciences have heavily modeled themselves on the natural sciences, hoping to achieve scientific rigor and societal recognition through quantification. Yet the generalizability and reproducibility crises show that this path has clear limits. On the replication crisis in another article

This does not mean that qualitative methods are automatically superior. They also have significant limitations and remain embedded in the empirical paradigm. Perhaps the way forward lies in methods that do not reduce the social to measurability, but instead take its ambiguity and context-dependence seriously.

The central question remains: Should the social sciences continue to be forced to speak the language of numbers—or dare we imagine a science that takes the social seriously, beyond empiricism?

References 

Borsboom, D., Mellenbergh, G. J., & van Heerden, J. (2004). The concept of validity. Psychological Review, 111(4), 1061–1071. https://doi.org/10.1037/0033-295X.111.4.1061

Cartwright, N. (2007). Are RCTs the gold standard? BioSocieties, 2(1), 11–20. https://doi.org/10.1017/S1745855207005020

Cronbach, L. J., & Meehl, P. E. (1955). Construct validity in psychological tests. Psychological Bulletin, 52(4), 281–302. https://doi.org/10.1037/h0040957

Deaton, A. (2010). Instruments, randomization, and learning about development. Journal of Economic Literature, 48(2), 424–455. https://doi.org/10.1257/jel.48.2.424

Deaton, A., & Cartwright, N. (2018). Understanding and misunderstanding randomized controlled trials. Social Science & Medicine, 210, 2–21. https://doi.org/10.1016/j.socscimed.2017.12.005

Duflo, E. (2017). The economist as plumber. American Economic Review, 107(5), 1–26. https://doi.org/10.1257/aer.p20171153

Duflo, E., Banerjee, A. V., & Kremer, M. (2011). Poor Economics: A Radical Rethinking of the Way to Fight Global Poverty. PublicAffairs. 

Firestein, S. (2012). Ignorance: How It Drives Science. Oxford University Press.

Gigerenzer, G. (2004). Mindless statistics. The Journal of Socio-Economics, 33(5), 587–606. https://doi.org/10.1016/j.socec.2004.09.033

Hacking, I. (1999). The Social Construction of What? Harvard University Press.

Henrich, J., Heine, S. J., & Norenzayan, A. (2010). The weirdest people in the world? Behavioral and Brain Sciences, 33(2–3), 61–83. https://doi.org/10.1017/S0140525X0999152X

IRD le Mag’ (2021). Randomized trials in development economics: Revolution or illusion? https://lemag.ird.fr/en/randomized-trials-development-economics-revolution-or-illusion

Ioannidis, J. P. A. (2005). Why most published research findings are false. PLoS Medicine, 2(8), e124. https://doi.org/10.1371/journal.pmed.0020124

The Global Policy Journal (2019, October 15). The Randomistas just won the Nobel Economics prize. Here’s why RCTs aren’t a magic bullet. https://www.globalpolicyjournal.com/blog/15/10/2019/randomistas-just-won-nobel-economics-prize-heres-why-rcts-arent-magic-bullet

The Wire (2019, October 14). Explainer: What Banerjee, Duflo and Kremer Won the Economics Nobel For. https://thewire.in/economy/abhijit-banerjee-esther-duflo-michael-kremer-economics-nobel-prize

Yarkoni, T. (2022). The generalizability crisis. Behavioral and Brain Sciences, 45, e1. https://doi.org/10.1017/S0140525X20001685

Inspired by HBS Puar 
Authored by Rebekka Brandt