Frank - stock.adobe.com

Why we should all be worried about the AI bubble

The phrase 'don't believe the hype' has never been more apt - there are growing warnings of an AI investment bubble that could affect everyone if it bursts

It's becoming clearer that we are in a perilous financial situation globally. Fears over an “AI bubble” are being cited by the Bank of England, the International Monetary Fund and the boss of JP Morgan, Jamie Dimon.

If you want a sense of how insane the narrative is around AI investments, consider this: Thinking Machines Lab, an AI startup, recently raised $2bn funding on a valuation of $10bn.

The company has zero products, zero customers and zero revenues. The only thing it made public to its investors was the resume of its founder, Mira Murati, formerly chief technology officer at OpenAI. If that’s not hubris meeting market exuberance, what is?

But narrative is crucial here because it’s what’s driving all this insane investment in the future of AI or so-called artificial general intelligence (AGI), and it's important to examine which narrative you believe in if you are to protect yourself for what’s to come.

If I were to pick between the views of a politician such as UK prime minister Keir Starmer, and a writer such as Cory Doctorow, I’d put my bet on Doctorow. Contrast these two statements and see which you feel more comfortable with…

Starmer: “Today’s plan mainlines AI into the veins of this enterprising nation”.

Doctorow: “AI is the asbestos we are shovelling into the walls of our society and our descendants will be digging it out for generations”.

Puncture the bubble

Doctorow suggests the AI bubble needs to be punctured as soon as possible to “halt this before it progresses any further and to head off the accumulation of social and economic debt”.

He suggests doing that by taking aim at the basis for the AI bubble – namely, “creating a growth story by claiming that AI can do your job”.

AI is the asbestos we are shovelling into the walls of our society and our descendants will be digging it out for generations
Cory Doctorow

Claims about jobs disappearing to AI have been around since 2019 with Sam Altman, then leader of venture capital (VC) fund Y Combinator, speaking about radiology jobs disappearing in the future: “Human radiologists are already much worse than computer radiologists. If I had to pick a human or an AI to read my scan, I’d pick the AI.”

Fast forward six years to 2025 and look how that worked out. According to a recent report by Works in Progress, despite the fact that radiology combines digital images, clear benchmarks and repeatable tasks, demand for human radiologists is at an all-time high.

The report authors’ conclusions drive a horse and cart through the current AI/AGI narrative that if left unstopped will cause severe global economic pain: “In many jobs, tasks are diverse, stakes are high, and demand is elastic. When this is the case, we should expect software to initially lead to more human work, not less. The lesson from a decade of radiology models is neither optimism about increased output nor dread about replacement. Models can lift productivity, but their implementation depends on behaviour, institutions and incentives. For now, the paradox has held - the better the machines, the busier radiologists have become.”

Across other sectors too, the mythology around job losses is slowly being interrogated – for example, Yale University Budget Lab found no discernible disruption to labour markets since ChatGPT’s release 33 months ago.

The research goes on to state: “While this finding may contradict the most alarming headlines, it is not surprising given past precedents. Historically, widespread technological disruption in workplaces tends to occur over decades, rather than months or years. Computers didn’t become commonplace in offices until nearly a decade after their release to the public, and it took even longer for them to transform office workflows. Even if new AI technologies will go on to impact the labour market as much, or more dramatically, it is reasonable to expect that widespread effects will take longer than 33 months to materialise”.

Normal technology

In other words, AI is just, well, technology as we have always known it - or as experts Aryind Narayanan and Sayash Kapoor call AI, just “normal technology”. 

Importantly in their paper, AI as normal technology - An alternative to the vision of AI as a potential superintelligence, they identify key lessons from past technological revolutions - the slow and uncertain nature of technology adoption and diffusion; continuity between the past and future trajectory of AI in terms of social impact; and the role of institutions in shaping this trajectory. They also “strongly disagree with the characterisation of generative AI adoption as rapid, which reinforces our assumption about the similarity of AI diffusion to past technologies”

A good example of AI as normal technology without all the hype, hyperbole and billion-dollar burn rate, is the City of Austin, Texas. Here, an on-premise AI system helped the local government process building permits in days instead of months.

According to David Stout, CEO of WebAI, this was done “with no spectacle. No headlines. Just efficiency gains that will outlast the market cycle. He said, “That’s the point too often missed in the frenzy. Mega-models attract headlines, consume billions in capital, and struggle to demonstrate sustainable economics. Meanwhile, smaller, domain-specific systems are already delivering efficiency gains, cost savings and productivity improvements. The smart play isn’t to abandon AI, but to pivot towards models and deployments that will endure”.

Technology like we have always known it to be - not the insane fantasy of “superintelligence” that is powering this dangerous bubble.

The question to ask is, given the prediction of at least a 33-month lag before any return on investment, however small, will the markets wait another 33 months for their returns to materialise?

Protracted crisis

A recent report on MarketWatch suggests the AI bubble is now ”seventeen times the size of the dot com frenzy and four times the sub-prime bubble”. MarketWatch quotes financial analyst Julien Garran, who previously led UBS’s commodities strategy team, who said “AI now accounts for over four times the wealth trapped in the 2008 sub-prime mortgage bubble, which resulted in years of protracted crisis across the globe”.

Warnings from the Bank of England in its semi-annual Financial Policy Committee report are equally stark: “Uncertainty around the global risk environment increases the risk that markets have not fully priced in possible adverse outcomes, and a sudden correction could occur should any of these risks crystallize."

The bank also warned of "the risk of a sharp market for global financial markets amid AI bubble risks and political pressure on the Federal Reserve."

What a sudden correction means is that a collapse of the AI investment bubble will take trillions of investment with it, impacting us all.

Even more worrying is the issue of debt financing among those competing in the AI race – that is, all the tech bros. It now appears, according to Axios, that these companies are turning to private debt markets and special purpose vehicles for cash, which means this kind of borrowing does not have to show on their balance sheets.

Meta, for example, recently sought $29bn from private capital firms for its AI datacentres. This off-book debt financing should ring more alarm bells that something is terribly wrong with the AI growth narrative.

After all, as pointed out by the Axios analysts, “If hugely profitable tech companies need to mask their borrowings to fund AI spending, it signals they’re not confident that they’ll soon get the returns needed to justify such investments. That suggests the very spending powering today’s earnings boom can’t last forever.”

Unit economics

To go back to Cory Doctorow’s argument, we are not in the early days of the web, or Amazon, or other dot com companies that lost money before becoming profitable: “Those were all propositions with excellent unit economics. They got cheaper with every successive technological generation and the more customers they added, the more profitable they became”.

AI companies do not have excellent unit economics - in fact they have the opposite, according to Doctorow: “Each generation of AI has been vastly more expensive than the previous one, and each new AI customer makes the AI companies lose more money”.

[Only] about 5% of tasks will be able to be profitably performed by AI within 10 years
Daron Acemoglu

And if that's not sobering enough for the VC and private equity firms, then the circular investing going on between these tech firms should be a huge concern.

Microsoft is investing $10bn in OpenAI by giving free access to its servers. OpenAI reports this as an “investment,” then redeems these tokens at Microsoft datacentres, which Microsoft books as $10bn in revenue.

Bain & Co says the only way to make today's AI investments profitable “is for the sector to bring in $2tn by 2030,” which, according to the Wall Street Journal, is more than the revenue of Amazon, Google, Microsoft, Apple, Nvidia and Meta - combined.

Taking a closer look at US economic growth is surely more cause for concern.

According to Harvard economist Jason Furman’s analysis, GDP growth in the first half of 2025 was driven almost entirely by investment in information processing equipment and software. This spending was largely tied to the rapid expansion of AI infrastructure and datacentres.

While these tech sectors only made up 4% of total GDP, they contributed a staggering 92% of growth. Absent this investment, Furman estimates US GDP growth would have hovered around 0.1% on an annualised basis - barely above zero.

There is a lot riding on a technology that’s supposed to be godlike and all powerful but which, according to MIT Institute professor Daron Acemoglu, is far less likely to achieve the insane hyperbolic claims being made by the tech bros in an effort to win an unwinnable race.

Acemoglu estimates the 10-year effect of AI in the US will be that only “about 5% of tasks will be able to be profitably performed by AI within that timeframe,” with the GDP boost likely to be closer to 1% over that timespan. If that’s not a recipe for stock market collapse, what is?

Emperor's new clothes

Going back to the AI booster narrative and how it’s driving things, Doctorow is again incisive: “The most important thing about AI isn’t its technical capabilities or limitations. The most important thing is the investor story and the ensuing mania that has teed up an economic catastrophe that will harm hundreds of millions or even billions of people. AI isn't going to wake up, become super intelligent and turn you into paperclips - but rich people with AI investor psychosis are almost certainly going to make you much, much poorer”.

I'm not an economist, so I did what we are all supposed to do now for our enlightenment. I gave the machines built by the tech bros all the same prompt: “What fable best encapsulates the current AI bubble?”

Gemini, Perplexity and ChatGPT were all in agreement with nearly the same explanation of why they all picked the same story: "The emperor's new clothes remains the best classic fable to explain the AI bubble, as it encapsulates the collective willingness to believe in - and profit from - an imagined reality, until facts and external shocks eventually break the spell.”

Read more about the AI bubble

Read more on Artificial intelligence, automation and robotics