Stochastic parrotology

Telegraph octopuses, observational world models, conduit metaphors and the limits of reasoning from pre-training

2025-08-24 — 2025-12-11

Wherein the limits of language posed by transformers are examined, and the claim that models’ outputs are echoed from their training corpora without human understanding is set forth while training‑data provenance is noted.

adaptive
agents
AI safety
Bayes
bounded compute
semantics
cooperation
extended self
generative
grammar
language
learning
machine learning
meta learning
mind
model selection
networks
neural nets
NLP
optimization
probability
statmech
stringology
utility
when to compute
Figure 1: Go on, buy the sticker

There is some kind of thought-experiment fight triggered by the pre-training heavy life-cycle of LLMs. This had led to two different families of vibes about artificial intelligence. On one side, we have the Stochastic Parrots thesis: the idea that no matter how fluent a model becomes, it is fundamentally just mimicking the statistical properties of its training data—a “blurry JPEG” of the web, forever trapped in the Chinese Room of syntax without semantics. On the other side is the growing empirical reality of models that seem to reason, code, and strategize in ways that defy simple statistical explanation.

We could formalise this and make it precise, and possibly resolve the thesis (several theses) that are wrapped up here. If we did so, it would please me to call field of inquiry Stochastic Parrotology: the rigorous study of what can (and cannot) be learned from pre-training on observational data alone.

Recent history of the debate: As Emily Bender and Timnit Gebru argued in their seminal paper (Bender et al. 2021), without communicative intent or grounding in the physical world, these models may be nothing more than haphazard stitchers of linguistic form—parrots that don’t know what they are saying. This echoes John Searle’s famous “Chinese Room” argument: manipulating symbols correctly is not the same as understanding them.

Jacob Browning and Yann LeCun reinforce this in AI And The Limits Of Language, arguing that language is a low-bandwidth, lossy compression of human experience. In their view, “shallow understanding” is the ceiling for any system trained on text alone because it lacks the non-linguistic “know-how” the muscle memory of navigating physical reality—that grounds human intelligence. To them, an LLM is a mirror: it reflects our depth back at us, but is itself only a centimeter thick.

But there is a counter-current suggesting that “parroting” the structure of language might force a model to learn the structure of the world. In Blurry JPEG or Frozen Concentrate, Lance Fortnow (riffing on Ted Chiang) suggests that what these models are doing is approximating the Kolmogorov Structure Function—finding the smallest program that could generate the set of all possible human articles. If the most efficient way to compress the data of the world is to learn the rules of the world, then “mere prediction” becomes a path to understanding.

Figure 2

Similarly, Deontologistics argues in On Post-Searlean Critiques of LLMs that we shouldn’t be so quick to dismiss the “game” of language. Drawing on inferentialism, they suggest that being “entangled” in the social web of giving and asking for reasons might be enough to constitute a form of meaning, even without a body. If a model can reliably play the “Inaccessible Game” of information exchange—a concept explored by Neil Lawrence in Perpetual Motion, Superintelligence and the Inaccessible Game—it may be developing internal structures that are functionally indistinguishable from world models.

So, are we looking at a Benderian Octopus—a creature that can trick us until a bear attacks and it fails to hand us a stick? Are we witnessing the emergence of a Conduit (to use Michael Reddy’s metaphor) that has somehow become aware of the content it carries?

To move beyond vibes, we need to operationalize these debates. We must ask specific, falsifiable questions about what happens inside the black box during pre-training.

If we strip away the anthropomorphism, the debate seems to be about at least three things that I can discern. Can we mathematically and empirically demonstrate whether the following can arise purely from observing the text of others?

  1. Can models learn intervention distributions from pre-training?
  2. Can models learn world models from pre-training?
  3. Can models learn agency from pre-training?

Let’s unpack these.

1 Can models learn intervention distributions from pre-training?

  • The Skeptic: No. As Pearl would argue, you cannot derive causation (\(do(x)\)) from observation (\(see(x)\)) without assumptions.
  • The Parrotologist: Does the massive redundancy of text contain “natural experiments” that leak causal structure into the observational distribution?

Speaking as a guy who studies causation, I would say there is no slam dunk to be had here. Judea Pearl taught us that the answer about whether we can go from observation to intervention is “it depends” specifically, on the causal graph structure and the assumptions we are willing to make.

2 Can models learn world models from pre-training?

  • The Skeptic: No. They learn a map of language, not a map of the territory (Browning/LeCun).
  • The Parrotologist: Is the “Kolmogorov Structure Function” of the internet actually a simulation of the world that generated it? Does compression force the emergence of a simulator?

cf (Hao et al. 2023; Hu and Shu 2023; Richens and Everitt 2024; Richens et al. 2025; Wong et al. 2023; Yildirim and Paul 2024)

3 Can models learn agency from pre-training?

  • The Skeptic: No. Agency requires goals, temporal continuity, and feedback loops with an environment.
  • The Parrotologist: Can the “simulacra” of agents (characters in stories, arguments in essays) be appropriated by the model to form a coherent, goal-directed persona when prompted?

cf (Hao et al. 2023; Hu and Shu 2023; Richens and Everitt 2024; Richens et al. 2025; Yildirim and Paul 2024)

4 References

Bender, Gebru, McMillan-Major, et al. 2021. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜.” In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency.
Bender, and Koller. 2020. Climbing Towards NLU: On Meaning, Form, and Understanding in the Age of Data.” In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
Fedorenko, Piantadosi, and Gibson. 2024. Language Is Primarily a Tool for Communication Rather Than Thought.” Nature.
Francis, and Wonham. 1976. The Internal Model Principle of Control Theory.” Automatica.
Hao, Gu, Ma, et al. 2023. Reasoning with Language Model Is Planning with World Model.”
Hu, and Shu. 2023. Language Models, Agent Models, and World Models: The LAW for Machine Reasoning and Planning.”
Piantadosi, and Hill. 2022. Meaning Without Reference in Large Language Models.”
Raji, Bender, Paullada, et al. 2021. AI and the Everything in the Whole Wide World Benchmark.”
Richens, Abel, Bellot, et al. 2025. General Agents Contain World Models.” In ICML.
Richens, and Everitt. 2024. Robust Agents Learn Causal World Models.”
Wong, Grand, Lew, et al. 2023. From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought.”
Yildirim, and Paul. 2024. From Task Structures to World Models: What Do LLMs Know? Trends in Cognitive Sciences.