Stochastic parrotology

Telegraph octopuses, observational world models, conduit metaphors and the limits of reasoning from pre-training

2025-08-24 — 2025-12-11

Wherein the question of whether pre‑training yields causal inference is examined, and the possibility that models extract intervention distributions from text corpora and form simulators via a Kolmogorov‑like structure function is considered.

adaptive
agents
AI safety
Bayes
bounded compute
semantics
cooperation
extended self
generative
grammar
language
learning
machine learning
meta learning
mind
model selection
networks
neural nets
NLP
optimization
probability
statmech
stringology
utility
when to compute
Figure 1: Go on, buy the sticker

There’s a thought-experiment fight triggered by the pre-training-heavy life-cycle of LLMs. This has led to two different families of vibes about artificial intelligence. On one side, we have the Stochastic Parrots thesis: the idea that no matter how fluent a model becomes, it’s fundamentally just mimicking the statistical properties of its training data—a “blurry JPEG” of the web, forever trapped in the Chinese Room of syntax without semantics. On the other side is the growing empirical reality of models that seem to reason, code, and strategize in ways that defy simple statistical explanation.

We could formalize this, make it precise, and possibly resolve the theses wrapped up here. If we did so, it would please me to call the field of inquiry Stochastic Parrotology: the rigorous study of what can (and cannot) be learned from pre-training on observational data alone.

Recent history of the debate: As Emily Bender and Timnit Gebru argued in their seminal paper (Bender et al. 2021), without communicative intent or grounding in the physical world, these models may be nothing more than haphazard stitchers of linguistic form—parrots that don’t know what they are saying. This echoes John Searle’s famous “Chinese Room” argument: manipulating symbols correctly is not the same as understanding them.

Jacob Browning and Yann LeCun reinforce this in AI And The Limits Of Language, arguing that language is a low-bandwidth, lossy compression of human experience. In their view, “shallow understanding” is the ceiling for any system trained on text alone because it lacks the non-linguistic “know-how”—the muscle memory of navigating physical reality—that grounds human intelligence. To them, an LLM is a mirror: it reflects our depth back at us, but is itself only a centimetre thick.

But there is a counter-current suggesting that “parroting” the structure of language might force a model to learn the structure of the world. In Blurry JPEG or Frozen Concentrate, Lance Fortnow (riffing on Ted Chiang) suggests that what these models are doing is approximating the Kolmogorov Structure Function—finding the smallest program that could generate the set of all possible human articles. If the most efficient way to compress the data of the world is to learn the rules of the world, then “mere prediction” becomes a path to understanding.

Figure 2

Similarly, Deontologistics argues in On Post-Searlean Critiques of LLMs that we shouldn’t be so quick to dismiss the “game” of language. Drawing on inferentialism, they suggest that being “entangled” in the social web of giving and asking for reasons might be enough to constitute a form of meaning, even without a body. If a model can reliably play the “Inaccessible Game” of information exchange—a concept Neil Lawrence explores in Perpetual Motion, Superintelligence and the Inaccessible Game—it may develop internal structures functionally indistinguishable from world models.

So, are we looking at a Benderian Octopus—a creature that can trick us until a bear attacks and it fails to hand us a stick? Are we witnessing the emergence of a Conduit (to use Michael Reddy’s metaphor) that has somehow become aware of the content it carries?

To move beyond vibes, we need to operationalize these debates. We must ask specific, falsifiable questions about what happens inside the black box during pre-training.

If we strip away the anthropomorphism, the debate seems to be about at least three things that I can discern. Can we mathematically and empirically demonstrate whether the following can arise purely from observing others’ text?

  1. Can models learn intervention distributions from pre-training?
  2. Can models learn world models from pre-training?
  3. Can models learn agency from pre-training?

Let’s unpack these.

1 Can models learn intervention distributions from pre-training?

  • The Skeptic: No. As Pearl would argue, you can’t derive causation (\(do(x)\)) from observation (\(see(x)\)) without assumptions.
  • The Parrotologist: Does the massive redundancy of text contain “natural experiments” that leak causal structure into the observational distribution?

Speaking as someone who studies causation, I’d say there’s no slam dunk here. Judea Pearl taught us that whether we can go from observation to intervention “depends” — specifically on the causal graph structure and the assumptions we’re willing to make.

2 Can models learn world models from pre-training?

  • The Skeptic: No. They learn a map of language, not a map of the territory (Browning/LeCun).
  • The Parrotologist: Is the “Kolmogorov Structure Function” of the internet actually a simulation of the world that generated it? Does compression force the emergence of a simulator?

cf (Hao et al. 2023; Hu and Shu 2023; Richens and Everitt 2024; Richens et al. 2025; Wong et al. 2023; Yildirim and Paul 2024)

3 Can models learn agency from pre-training?

  • The Skeptic: No. Agency requires goals, temporal continuity, and feedback loops with an environment.
  • The Parrotologist: Can the model appropriate the “simulacra” of agents (characters in stories, arguments in essays) to form a coherent, goal-directed persona when prompted?

cf (Hao et al. 2023; Hu and Shu 2023; Richens and Everitt 2024; Richens et al. 2025; Yildirim and Paul 2024)

4 Can we learn to act from pre-recorded off-policy data?

The reinforcement learninig variant of this debate is whether an agent can learn to act effectively from a static dataset of observations and actions, without further interaction with the environment.

  • The Skeptic: No. Without on-policy experience, you can’t learn to act effectively.
  • The Parrotologist: Can we use pre-trained models as simulators to bootstrap policy learning?

5 References

Bender, Gebru, McMillan-Major, et al. 2021. On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜.” In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency.
Bender, and Koller. 2020. Climbing Towards NLU: On Meaning, Form, and Understanding in the Age of Data.” In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics.
Fedorenko, Piantadosi, and Gibson. 2024. Language Is Primarily a Tool for Communication Rather Than Thought.” Nature.
Francis, and Wonham. 1976. The Internal Model Principle of Control Theory.” Automatica.
Hao, Gu, Ma, et al. 2023. Reasoning with Language Model Is Planning with World Model.”
Hu, and Shu. 2023. Language Models, Agent Models, and World Models: The LAW for Machine Reasoning and Planning.”
Piantadosi, and Hill. 2022. Meaning Without Reference in Large Language Models.”
Pleyer, Perlman, Lupyan, et al. 2025. The ‘design features’ of language revisited.” Trends in Cognitive Sciences.
Raji, Bender, Paullada, et al. 2021. AI and the Everything in the Whole Wide World Benchmark.”
Richens, Abel, Bellot, et al. 2025. General Agents Contain World Models.” In ICML.
Richens, and Everitt. 2024. Robust Agents Learn Causal World Models.”
Wong, Grand, Lew, et al. 2023. From Word Models to World Models: Translating from Natural Language to the Probabilistic Language of Thought.”
Yildirim, and Paul. 2024. From Task Structures to World Models: What Do LLMs Know? Trends in Cognitive Sciences.