Epistemic communities

Descriptive and normative

2021-08-24 — 2026-04-17

Wherein the several aims of collective truth-seeking are enumerated, and their mutual incompatibility is noted, with peer review instanced as the community most amenable to experimental manipulation.

adaptive
agents
collective knowledge
communicating
cooperation
democracy
distributed
economics
game theory
how do science
incentive mechanisms
institutions
networks
provenance
sociology
standards
virality

There is a lot to say about epistemic communities in the abstract. Will I ever get around to saying it in full, or will I defer my effort in favour of specific cases? Let us find out.

What do I even mean by “epistemic community”? I have in mind any collective that produces and maintains shared knowledge, whether that’s a scientific discipline, a news media ecosystem, or a Wikipedia subcommunity. The common thread is that they are all trying to figure out what’s true, and they have some kind of social structure that they hope to leverage to help them do it.

The science community has thought hard about this; journalism seems to have thought about it somewhat. I think it is becoming relevant in new ways in online communities and civic tech.

Figure 1

1 What is a good epistemic?

We can ask an epistemic community to do any of the following:

  • converge on true beliefs at the margin, as fast as possible
  • maintain enough diversity that the community does not lock in on a confident falsehood
  • produce consensus-backed action (e.g. public-health messaging)
  • let individuals skim useful knowledge cheaply
  • ratify the status of its members

These aims trade off. A community optimised for fast convergence will purge heterodoxy we might later need (O’Connor, Goldberg, and Goldman 2024; Weatherall, O’Connor, and Bruner 2018). One optimised for diversity will struggle to close questions that actually are closed. One that grades its members on status will find its incentives drifting off the object level.

So the first design question is: which of these are we asking of this particular community, under what budget? IMO a lot of debate about “trust in experts” is at bottom a debate about which of these we think experts are for, which is why the participants talk past each other.

2 Bayesian epistemics

An inverse design approach: how should we design a system to optimise for truthfulness? See Bayesian epistemics. The peer-prediction and Bayesian Truth Serum family — which also appears in peer review — is the sharp end of this: we can elicit truthful reports without ground truth, at least in principle.

3 Formal models of the enterprise

Bayesian epistemics gives the micro-foundations; an adjacent literature takes the next step and asks what happens once we wire scoring rules into the principal-agent problems that actually constitute a scientific community.

Chen et al. (2023) combine proper scoring rules with the principal-agent model: under mild conditions, a principal can learn to incentivise accurate information acquisition by paying agents according to a scoring rule on their reported posteriors. This connects the Bayesian epistemics toolkit to the design of research grants — it is at least thinkable that a funding agency could run something like a scoring-rule market and do better than “did the results look plausible to a panel”.

The paper-reader relationship is also principal-agent, but messier, because the reader is not the principal who paid the author — they are paying only with their attention. Wei (2021) models this: when the receiver has to pay a cost to process a signal (the realistic case for scientific literature), the sender has latitude to persuade rather than inform, because the receiver cannot afford to audit everything. This formalises the worry that a lot of what looks like science is in fact rhetoric aimed at busy reviewers, and it tells us when the failure mode bites hardest — high reader-cost plus high author-reputation-return equals a persuasion equilibrium.

Aggregation can fail in its own way. Kong and Schoenebeck (2022) study what happens when individual agents’ signals are correlated and the agents do not know how correlated they are: the aggregate can be confidently wrong. Prediction-market prices, under their model, can sustain false consensus for longer than one might hope. A useful corrective if we were hoping prediction markets would simply solve the problem for us.

Mann and Helbing (2017) ask what incentive scheme maximises the diagnostic accuracy of a committee that has to decide, from its members’ private signals, what is true. The minimum-variance unbiased answer is not “majority vote with equal weights”; it depends on the correlation structure of the signals and on how we reward members for reports that turn out to be right versus wrong. The optimal reward scheme pays disproportionately for dissent that turns out to be correct — formal backing for the diversity-of-practice argument, and a quantitative counterpart to the bridging-based-ranking intuition below.

4 The deference economy

A community that produces knowledge has to decide whose claims to defer to. It seems to me that most “epistemic crises” in public debate actually play out here — not on the object level, but on the meta level of which sources carry weight and why.

Two relevant notebooks:

  • Status draws the standard distinction between dominance and prestige. Prestige — deference voluntarily conferred toward competence — is the ingredient that makes scientific cooperation scale beyond kin.
  • Reputation systems surveys attempts to formalize prestige into something a machine can compute.

The usual failure mode is that the prestige signal decouples from the underlying competence. Dan Williams (Williams 2022, 2023) has a long-running argument that many apparently-crazy beliefs are ingroup signalling dressed up as epistemics. That framing lines up with invasive arguments and with Timur Kuran’s preference-falsification cascades1. If the reward for expressing belief B is uncorrelated with B being true, we should expect B to drift.

Science-studies has formal models of this. Cailin O’Connor and James Weatherall’s agent-based network models (O’Connor and Weatherall 2017; O’Connor 2024; Weatherall, O’Connor, and Bruner 2018) show how an innocently-homophilous community can lock in false consensus. Paul Thagard’s coherence-based work (Thagard 1993, 2005) covers related ground from a cognitive-science angle. David Wolpert’s stochastic/thermodynamic analyses (Wolpert and Kinney 2023, 2024; Wolpert and Harper 2025) are on my list but I haven’t cracked them yet. I should probably work through the Seselja survey (Šešelja 2022) before forming strong opinions here.

5 Peer review

Peer review is the most-instrumented epistemic community we have. People bolt on incentive mechanisms and watch what happens; they A/B test on NeurIPS. Some fuzzy lessons learned:

  • Admission control. We get more effort out of reviewers if reviewing is priced into the right to submit. This is a baby version of the general problem: how do we stop free-riding on the commons of attention?
  • Randomness against manipulation. Deterministic assignment is gameable; controlled randomness buys robustness and cheap experiments.
  • Aggregating miscalibrated judges. People use 1–10 scales differently. Simple averaging bakes the noise in. Social choice theory has something to say about this, as do empirical calibration methods.
  • Elicitation with proper scoring rules. Asking reviewers to forecast observable outcomes, then scoring those forecasts, is incentive-compatible in a way that rubric ratings are not.

All four of these apply to any epistemic community that has to aggregate distributed expert judgement — editorial boards, Wikipedia’s admin layer, Q&A sites, Community Notes, grand juries, even deliberative-poll experiments. TODO: write this comparison out properly.

A subtler lesson from the NeurIPS experiments: the process itself is the experimental subject. If we never A/B test our community’s rules, we cannot learn. A lot of institutions are “unfalsifiable” in this sense, which is a design mistake, not a neutrality.

6 AI at the margins

LLMs change the economics in ways I have not finished thinking through. The cases I find most concrete:

  • Automated criticism. AI-assisted review for scientific papers (Allen-Zhu and Xu 2025); the Black Spatula Project; o1 catching arithmetic errors; roastmypost. Arbitration — analysing an existing dispute — is cognitively easier than generating a novel critique, which suggests a division of labour where LLMs mop up the mechanical-correctness pass and humans focus on taste and framing.
  • Bridging algorithms. Community Notes uses a matrix-factorisation ranker that surfaces notes rated helpful by users who usually disagree; Polis and vTaiwan apply a related pattern to structured deliberation. See bridging-based ranking for the mathematics and the broader family (Ovadya & Thorburn’s writeups, the Forethought design sketches for collective epistemics).
  • Deliberative tooling. Polis / vTaiwan provide structured, AI-assisted group deliberation. The DeepMind “Habermas Machine” experiment2 is in the same family.
  • Degraded signals. An LLM-saturated Quora is worse than the pre-LLM Quora. The cheaper passable-looking text gets, the more expensive it becomes to sort a community from its doppelgängers. Related: spamularity, misinformation, AI persuasion.

I suspect “does AI help or hurt?” is the wrong question — it’s underspecified. Better: how does AI shift the cost structure of the underlying mechanism? A community whose signal was mostly “could this person write a passable essay” is in trouble; one whose signal was mostly “could this person stake their reputation on a concrete forecast” is not.

Figure 2

7 Paying for truth

Internal review mechanisms are only half of the problem; the other half is the external payoff. If the reward for producing engaging slop is higher than the reward for producing truthful content, truth loses at the margin no matter how well the internal mechanism is designed. Public sphere business models § Paying for truth walks through the candidate funders (advertising, subscriptions, patronage, public broadcasting, philanthropy, prediction markets, reputation economies) and their characteristic failure modes.

To a first approximation, truth is a public good whose quality is only legible in the long run, which puts the problem close to “how do we fund basic research” — old and only partially solved.

8 Grab bag of design choices

Common interesting design themes from the above:

  1. Entry and exit costs — admission control, review credits, Sybil resistance (see reputation systems).
  2. Scale and context — Dunbar-number limits, subgroup granularity, context collapse. See Gordon Brander on thinking together and common knowledge.
  3. Anonymity vs pseudonymity vs identified. The peer-review trade-offs generalise.
  4. Elicitation format — freeform text, numeric rubric, probabilistic forecast, peer-prediction report.
  5. Aggregation rule — simple average, calibrated average, matrix-factorised bridging, prediction-market price, jury.
  6. Feedback loops — do members see outcomes? does the community know its own calibration?
  7. Adversarial robustness — collusion rings, cycles, fake identities, bot swarms, LLM spam.
  8. Funding source, and how it colours the above.

Most interesting design work happens at the intersection of two or three of these.

9 Research I should name-check

In no particular order.

  • Cailin O’Connor and James Weatherall, The Misinformation Age, plus the underlying network-of-scientists simulations (O’Connor and Weatherall 2017; O’Connor 2024; O’Connor, Goldberg, and Goldman 2024; Weatherall, O’Connor, and Bruner 2018). In-bib; need to actually read.
  • Jonathan Rauch, The Constitution of Knowledge. Popular rather than technical; useful for the framing of “reality-based community” as a social institution with rules. (Rauch 2021)]
  • Hugo Mercier and Dan Sperber, The Enigma of Reason. The argumentative theory of reasoning implies that epistemic communities, not individuals, are the unit of cognition. (Mercier and Sperber 2017)
  • Dan Williams, Conspicuous Cognition. Blog aside, his journal work on epistemic signalling (Williams 2021a, 2022, 2023) is already in-bib.
  • Kevin Zollman on scientific network topology and the epistemic value of transient disagreement (the “Zollman effect”). (Kevin J. S. Zollman 2007; Kevin JS. Zollman 2012)
  • Henry Farrell and Cosma Shalizi, Cognitive Democracy, for the analytical frame. Farrell’s later work on AI-era epistemics (Farrell, Mercier, and Schwartzberg 2022) is in-bib.
  • Timur Kuran, Private Truths, Public Lies (preference falsification). Sits naturally next to Williams. (Kuran 1997)
  • Frischmann, Madison, and Strandburg, Governing Knowledge Commons (Cambridge). Ostromian analysis applied specifically to knowledge production. (Frischmann, Madison, and Strandburg 2014)
  • Elinor Ostrom, Governing the Commons. Already arrives via public sphere business models; probably needs its own entry point here.
  • Aviv Ovadya and Luke Thorburn on bridging-based ranking and “platform democracy”. TODO cite.
  • Renée DiResta, Invisible Rulers. Influence operations as an industry. TODO cite.
  • Chris Bail, Breaking the Social Media Prism. Empirical work on echo chambers where the results are not the obvious ones (cf. (Eady et al. 2023)).
  • J. Nathan Matias and the Citizens and Technology Lab. Experimental manipulation of online-community norms.
  • David Wolpert, stochastic/thermodynamic analyses of epistemic communities (Wolpert and Kinney 2023, 2024; Wolpert and Harper 2025).
  • Paul Thagard on coherentist social epistemology (Thagard 1993, 2005).
  • Seselja et al. (Šešelja 2022), a survey of agent-based models of epistemic communities. Read this before the originals.
  • Community Notes research program (Birdwatch / Wojcik et al.). TODO cite.
  • Audrey Tang, Polis, and vTaiwan writeups. TODO cite.
  • Jason McKenzie Alexander on the norms of science and evolutionary game theory of epistemic communities. TODO cite.
  • Joshua Habgood-Coote on whether “misinformation” is a well-posed category at all. TODO cite.

10 See also

11 Incoming

News media and the public’s shared reality. Fake news, indeterminate news, incomplete news, alternative facts, strategic inference, kompromat, agnotology, facebooking to a molecular level. Basic media literacy, and whether it helps. As seen in elections and provocateur twitter bots.

Samuel Butler:

The public buys its opinions as it buys its meat, or takes in its milk, on the principle that it is cheaper to do this than to keep a cow. So it is, but the milk is more likely to be watered.

12 References

Allen-Zhu, and Xu. 2025. DOGE: Reforming AI Conferences and Towards a Future Civilization of Fairness and Justice.” SSRN Scholarly Paper.
Barlev, and Neuberg. 2025. Rational reasons for irrational beliefs.” The American Psychologist.
Bazzoli. 2022. Open Science and Epistemic Pluralism: A Tale of Many Perils and Some Opportunities.” Industrial and Organizational Psychology.
Berger, and Luckmann. 2016. “The Social Construction of Reality.” In Social Theory Re-Wired.
Bhattacharya, and Packalen. 2020. Stagnation and Scientific Incentives.” Working Paper 26752. Working Paper Series.
Bolton, and Harris. 1999. Strategic Experimentation.” Econometrica.
Channing, and Ghosh. 2025. AI for Scientific Discovery Is a Social Problem.”
Chen, Wu, Wu, et al. 2023. Learning to Incentivize Information Acquisition: Proper Scoring Rules Meet Principal-Agent Model.” In Proceedings of the 40th International Conference on Machine Learning.
Cross. 2013. Rethinking Epistemic Communities Twenty Years Later.” Review of International Studies.
Eady, Paskhalis, Zilinsky, et al. 2023. Exposure to the Russian Internet Research Agency Foreign Influence Campaign on Twitter in the 2016 US Election and Its Relationship to Attitudes and Voting Behavior.” Nature Communications.
Falandays, Kaaronen, Moser, et al. 2022. All Intelligence Is Collective Intelligence.”
Farrell, Mercier, and Schwartzberg. 2022. Analytical Democratic Theory: A Microfoundational Approach.” American Political Science Review.
Ferraz-Caetano. 2025. Modeling Innovations: Levels of Complexity in the Discovery of Novel Scientific Methods.” Philosophies.
Flore. 2025. Synthetic Friends: AI Companions and the Future of Disinformation.” SSRN Scholarly Paper.
Frischmann, Madison, and Strandburg. 2014. Governing Knowledge Commons.”
Halpern, and Moses. 1990. Knowledge and Common Knowledge in a Distributed Environment.” Journal of the ACM (JACM).
Hanson, Barreiro, Crosetto, et al. 2024. The Strain on Scientific Publishing.” Quantitative Science Studies.
Heesen, and Journal of Philosophy Inc. 2018. Why the Reward Structure of Science Makes Reproducibility Problems Inevitable: The Journal of Philosophy.
Hoelzemann, and Klein. 2021. Bandits in the Lab.” Quantitative Economics.
Hsiao, Lin, Tang, et al. 2018. vTaiwan: An Empirical Study of Open Consultation Process in Taiwan.”
Kaelbling, Littman, and Moore. 1996. Reinforcement Learning: A Survey.” Journal of Artifical Intelligence Research.
Kong, and Schoenebeck. 2022. False Consensus, Information Theory, and Prediction Markets.”
Krumrei-Mancuso, Pärnamets, Bland, et al. 2025. Toward an Understanding of Collective Intellectual Humility.” Trends in Cognitive Sciences.
Kuran. 1997. Private Truths, Public Lies: The Social Consequences of Preference Falsification.
Littman. 2021. Collusion Rings Threaten the Integrity of Computer Science Research.” Communications of the ACM.
Madison, Frischmann, and Strandburg. 2018. Knowledge Commons.” SSRN Scholarly Paper.
Mann, and Helbing. 2017. Optimal Incentives for Collective Intelligence.” Proceedings of the National Academy of Sciences.
McElreath, and Smaldino. 2015. Replication, Communication, and the Population Dynamics of Scientific Discovery.” PLOS ONE.
Mercier, and Sperber. 2017. The Enigma of Reason.
Michelini, Osorio, Houkes, et al. 2023. “Scientific Disagreements and the Diagnosticity of Evidence: How Too Much Data May Lead to Polarization.” Journal of Artificial Societies and Social Simulation.
Musen, O’Connor, Schultes, et al. 2022. Modeling Community Standards for Metadata as Templates Makes Data FAIR.” Scientific Data.
NATO Science and Technology Organization, Human Factors and Medicine Panel (HFM-ET-356). 2023. Mitigating and Responding to Cognitive Warfare.” Technical Report TR-HFM-ET-356.
Nelson. 1959. The Simple Economics of Basic Scientific Research.” Journal of Political Economy.
O’Connor. 2024. Modelling Scientific Communities.
O’Connor, Goldberg, and Goldman. 2024. Social Epistemology.” In The Stanford Encyclopedia of Philosophy.
O’Connor, and Weatherall. 2017. Scientific Polarization.” European Journal for Philosophy of Science.
———. 2021. Modeling How False Beliefs Spread.” In The Routledge Handbook of Political Epistemology.
Patterson, and Williams. 2005. Maintaining Research Traditions on Place: Diversity of Thought and Scientific Progress.” Journal of Environmental Psychology.
Pavitt. 1991. What Makes Basic Research Economically Useful? Research Policy.
Piekarski. 2021. Understanding Predictive Processing. A Review.” AVANT. Pismo Awangardy Filozoficzno-Naukowej.
Rauch. 2021. The Constitution of Knowledge: A Defense of Truth.
Ritchie. 2020. Science Fictions: How Fraud, Bias, Negligence, and Hype Undermine the Search for Truth.
Rubin, and Schneider. 2021. Priority and Privilege in Scientific Discovery.” Studies in History and Philosophy of Science Part A.
Šešelja. 2022. Agent‐based Models of Scientific Interaction.” Philosophy Compass.
Shin, Price, Wolpert, et al. 2020. Scale and Information-Processing Thresholds in Holocene Social Evolution.” Nature Communications.
Smaldino, and McElreath. 2016. The Natural Selection of Bad Science.” Royal Society Open Science.
Smith. 2022. The Internet Is Not What You Think It Is: A History, a Philosophy, a Warning.
Tessler, Bakker, Jarrett, et al. 2024. AI Can Help Humans Find Common Ground in Democratic Deliberation.” Science.
Thagard. 1993. “Societies of Minds: Science as Distributed Computing.” Studies in History and Philosophy of Modern Physics.
———. 1997. “Collaborative Knowledge.” Noûs.
———. 2005. “How to Be a Successful Scientist.” Scientific and Technological Thinking.
———. 2007. Coherence, Truth, and the Development of Scientific Knowledge.” Philosophy of Science.
The Constitution of Knowledge.” 2021. The Constitution of Knowledge.
Thorburn, and Ovadya. 2023. How to Redesign Social Media Algorithms to Bridge Divides.”
Weatherall, O’Connor, and Bruner. 2018. How to Beat Science and Influence People: Policy Makers and Propaganda in Epistemic Networks.”
Wei. 2021. Persuasion Under Costly Learning.” Journal of Mathematical Economics.
Williams. 2021a. Epistemic Irrationality in the Bayesian Brain.” British Journal for the Philosophy of Science.
———. 2021b. Socially Adaptive Belief.” Mind & Language.
———. 2021c. Motivated Ignorance, Rationality, and Democratic Politics.” Synthese.
———. 2022. Signalling, Commitment, and Strategic Absurdities.” Mind and Language.
———. 2023. The Case for Partisan Motivated Reasoning.”
Wolpert, and Harper. 2025. The Computational Power of a Human Society: A New Model of Social Evolution.”
Wolpert, and Kinney. 2023. Stochastic Mathematical Systems.”
———. 2024. A Stochastic Model of Mathematics and Science.” Foundations of Physics.
Yin, Dong, Wang, et al. 2021. Science as a Public Good: Public Use and Funding of Science.” Working Paper. Working Paper Series.
Zollman, Kevin J. S. 2007. The Communication Structure of Epistemic Communities.” Philosophy of Science.
———. 2009. The Epistemic Benefit of Transient Diversity.” Erkenntnis.
Zollman, Kevin JS. 2012. Social Network Structure and the Achievement of Consensus.” Politics, Philosophy & Economics.

Footnotes

  1. Kuran, Private Truths, Public Lies, and the Unanticipated Consequences of Preference Falsification. TODO cite.↩︎

  2. Tessari Prata et al., AI can help humans find common ground in democratic deliberation. TODO cite.↩︎