Economics of foundation models

Microeconomic of compute

March 23, 2023 — April 14, 2025

agents
bounded compute
collective knowledge
economics
edge computing
extended self
faster pussycat
incentive mechanisms
innovation
language
machine learning
neural nets
NLP
swarm
technology
UI
when to compute
Figure 1

1 Returns to scale for large AI firms

2 Spamularity, dark forest, textpocalypse

See Spamularity.

3 PR, hype, marketing

Figure 2

George Hosu, in a short aside, highlights the incredible marketing advantage of AI:

People that failed to lift a finger to integrate better-than-doctors or work-with-doctors supervised medical models for half a century are stoked at a chatbot being as good as an average doctor and can’t wait to get it to triage patients

The Tweet that Sank $100bn

Google’s Bard was undone on day two by an inaccurate response in the demo video where it suggested that the James Webb Space Telescope would take the first images of exoplanets.

This sounds like something the JWST would do but it’s not at all true.

So one tweet from an astrophysicist sank Alphabet’s value by 9%. This says a lot about how

  1. LLMs are like being at the pub with friends, it can say things that sound plausible and true enough, and no one really needs to check because who cares?

    Except we do because this is science, not a lads’ night out, and

  2. the insane speculative volatility of this AI bubble that the hype is so razor thin it can be undermined by a tweet with 44 likes.

I had a wonder if there’s any exploration of the ‘thickness’ of hype. Jack Stilgoe suggested looking at Borup et al. () which is evergreen but I feel like there’s something about the resilience of hype:

Like crypto was/is pretty thin in the scheme of things. High levels of hype but frenetic, unstable and quick to collapse.

AI has pretty consistent if pulsating hype gradually growing over the years while something like nuclear fusion is super thick (at least in the popular imagination) – remaining through decades of not-quite-ready and grasping the slightest indication of success.

I don’t know, if there’s nothing specifically on this, maybe I should write it one day.

Figure 3: Some of Tom Gauld’s caution signs

4 Empirical frontier models cash money costs

5 Democratisation of AI

A fascinating phenomenon..

6 Art and creativity

For now, see timeless works of art.

7 Data sovereignty

See data sovereignty.

8 AI tech soap opera

See AI tech as tragicomedy.

9 Incoming

I would prefer it as “AGI is less likely to abolish human labour value than you previously thought” rather than a blanket statement, but YMMV. * Ilya Sutskever: “Sequence to sequence learning with neural networks: what a decade” * Can the climate survive the insatiable energy demands of the AI arms race? * Why Quora isn’t useful anymore: A.I. came for the best site on the internet. * What Will Transformers Transform? – Rodney Brooks * Gradient Dissent, a list of reasons that large backpropagation-trained networks might be worrisome. Some interesting points in there, and some hyperbole. Also: If it were true that externalities come from backprop networks (i.e. that they are a kind of methodological pollution that produces private benefits but public costs) then what kind of mechanisms should disincentivise them? * C&C Against Predictive Optimisation * Stanford CRFM

In this post, we evaluate whether major foundation model providers currently comply with these draft requirements and find that they largely do not. Foundation model providers rarely disclose adequate information regarding the data, compute, and deployment of their models as well as the key characteristics of the models themselves. In particular, foundation model providers generally do not comply with draft requirements to describe the use of copyrighted training data, the hardware used and emissions produced in training, and how they evaluate and test models. As a result, we recommend that policymakers prioritise transparency, informed by the AI Act’s requirements. Our assessment demonstrates that it is currently feasible for foundation model providers to comply with the AI Act, and that disclosure related to foundation models’ development, use, and performance would improve transparency in the entire ecosystem.

10 References

Acemoglu, and Restrepo. 2018. Artificial Intelligence, Automation and Work.” In The Economics of Artificial Intelligence: An Agenda.
Akerlof, Holden, and Li. 2024. Getting the Picture.” SSRN Scholarly Paper.
Andrus, Dean, Gilbert, et al. 2021. AI Development for the Public Interest: From Abstraction Traps to Sociotechnical Risks.”
Babina, Fedyk, He, et al. 2021. Artificial Intelligence, Firm Growth, and Industry Concentration.” SSRN Scholarly Paper ID 3651052.
Barke, James, and Polikarpova. 2022. Grounded Copilot: How Programmers Interact with Code-Generating Models.”
Borup, Brown, Konrad, et al. 2006. The Sociology of Expectations in Science and Technology.” Technology Analysis & Strategic Management.
Bowman. 2023. Eight Things to Know about Large Language Models.”
Bullock, and Chen. 2024. The Brave New World of AI: Implications for Public Sector Agents, Organisations, and Governance.” Asia Pacific Journal of Public Administration.
Cheng, and McKernon. 2024. 2024 State of the AI Regulatory Landscape.”
Dahlin. 2022. Are Robots Really Stealing Our Jobs? Perception Versus Experience.” Socius.
Danaher. 2018. Toward an Ethics of AI Assistants: An Initial Framework.” Philosophy & Technology.
Dell, and Nestoriak. 2020. “Assessing the Impact of New Technologies on the Labor Market: Key Constructs, Gaps, and Data Collection Strategies for the Bureau of Labor Statistics.”
Douglas, and Verstyuk. 2025. Progress in Artificial Intelligence and Its Determinants.”
Farrell, Gopnik, Shalizi, et al. 2025. Large AI Models Are Cultural and Social Technologies.” Science.
Felten, Raj, and Seamans. 2019. The Occupational Impact of Artificial Intelligence: Labor, Skills, and Polarization.” SSRN Scholarly Paper ID 3368605.
Grimberg, and Mason. 2025. Building Proficiency in GAI: Key Competencies for Success.” Qeios.
Grossmann, Feinberg, Parker, et al. 2023. AI and the Transformation of Social Science Research.” Science.
Handa, Tamkin, McCain, et al. n.d. Which Economic Tasks Are Performed with AI? Evidence from Millions of Claude Conversations.”
Huang. 2024. The Labor Market Impact of Artificial Intelligence: Evidence from US Regions.” IMF Working Papers.
Kalyani, Bloom, Carvalho, et al. 2025. The Diffusion of New Technologies.” The Quarterly Journal of Economics.
Korinek. 2024. Economic Policy Challenges for the Age of AI.” Working Paper. Working Paper Series.
Kwa, West, Becker, et al. 2025. Measuring AI Ability to Complete Long Tasks.”
Lane, and Saint-Martin. 2021. The Impact of Artificial Intelligence on the Labour Market: What Do We Know so Far?
Lee. 2025. The Impact of Generative AI on Critical Thinking: Self-Reported Reductions in Cognitive Effort and Confidence Effects From a Survey of Knowledge Workers.”
Messeri, and Crockett. 2024. Artificial Intelligence and Illusions of Understanding in Scientific Research.” Nature.
Métraux. 1956. “A Steel Axe That Destroyed a Tribe, as an Anthropologist Sees It.” The UNESCO Courier: A Window Open on the World.
Naudé. 2022. The Future Economics of Artificial Intelligence: Mythical Agents, a Singleton and the Dark Forest.” IZA Discussion Papers, IZA Discussion Papers,.
Pelto. 1973. The snowmobile revolution: technology and social change in the Arctic.
Prettner, and Strulik. 2020. Innovation, Automation, and Inequality: Policy Challenges in the Race Against the Machine.” Journal of Monetary Economics.
Raman, Kumar Nair, Nedungadi, et al. 2024. Fake News Research Trends, Linkages to Generative Artificial Intelligence and Sustainable Development Goals.” Heliyon.
Shanahan. 2023. Talking About Large Language Models.”
Shumailov, Shumaylov, Zhao, et al. 2023. The Curse of Recursion: Training on Generated Data Makes Models Forget.”
Smith. 2022. The Internet Is Not What You Think It Is: A History, a Philosophy, a Warning.
Spector, Link to external site, and Ma. 2019. Inquiry and critical thinking skills for the next generation: from artificial intelligence back to human intelligence.” Smart Learning Environments.
Srivastava, and Bullock. 2024. AI, Global Governance, and Digital Sovereignty.”
Susskind, and Susskind. 2018. The Future of the Professions.” Proceedings of the American Philosophical Society.
Sytsma, and Sousa. 2023. Artificial Intelligence and the Labor Force: A Data-Driven Approach to Identifying Exposed Occupations.”
Wang, Chen, and Chen. 2024. How Artificial Intelligence Affects the Labour Force Employment Structure from the Perspective of Industrial Structure Optimisation.” Heliyon.
Zwetsloot, and Dafoe. 2019. Thinking About Risks From AI: Accidents, Misuse and Structure.” Lawfare.