Business The great AI delusion is falling apart - New research suggests the chorus of techno-optimism is based on falsehoods

Link

By Andrew Orlowski, The Telegraph
14 July 2025 11:00am BST


Is the secret of artificial intelligence that we have to kid ourselves, like an audience at a magic show?
Some fascinating new research suggests that self-deception plays a key role in whether AI is perceived to be a success or a dud.
In a randomised controlled trial – the first of its kind – experienced computer programmers could use AI tools to help them write code. What the trial revealed was a vast amount of self-deception.
“The results surprised us,” research lab METR reported. “Developers thought they were 20pc faster with AI tools, but they were actually 19pc slower when they had access to AI than when they didn’t.”

In reality, using AI made them less productive: they were wasting more time than they had gained. But what is so interesting is how they swore blind that the opposite was true.
If you think AI is helping you in your job, perhaps it’s because you want to believe that it works.
Since OpenAI’s ChatGPT was thrown open to the general public in late 2022, pundits have been forecasting huge productivity gains from deploying AI. They hope that it will supercharge growth and boost GDP. This has become the default opinion in high-status policy circles.
But all this techno-optimism is founded on delusion. The “lived experience” of using real tools in the real world paints a very different picture.

The past few days have felt like a turning point, as the reluctance of pointing out the emperor’s new clothes diminishes.
“I build AI agents for a living, it’s what I do for my clients,” wrote one Reddit user. “The gap between the hype and what’s actually happening on the ground is turning into a canyon”
AI isn’t reliable enough to do the job promised. According to an IBM survey of 2,000 chief executives, three out of four AI projects have failed to show a return on investment, which is a remarkably high failure rate.

Don’t hold your breath for a white-collar automation revolution either: AI agents fail to complete the job successfully about 65 to 70pc of the time, according to a study by Carnegie Mellon University and Salesforce.
The analyst firm Gartner Group has concluded that “current models do not have the maturity and agency to autonomously achieve complex business goals or follow nuanced instructions over time.” Gartner’s head of AI research Erick Brethenoux says: “AI is not doing its job today and should leave us alone”.
It’s no wonder that companies such as Klarna, which laid off staff in 2023 confidently declaring that AI could do their jobs, are hiring humans again.
This is extraordinary, and we can only have reached this point because of a historic self-delusion. People will even pledge their faith to AI working well despite their own subjective experience to the contrary, the AI critic Professor Gary Marcus noted last week.
“Recognising that it sucks in your own speciality, but imagining that it is somehow fabulous in domains you are less familiar with”, is something he calls “ChatGPT blindness”.

Much of the news is misleading. Firms are simply using AI as an excuse for retrenchment. Cost reduction is the big story in business at the moment.
Globally, President Trump’s erratic behaviour has induced caution, while in the UK, business confidence is at “historically depressed levels”, according to the Institute of Directors, reeling from Reeves’s autumn taxes. Attributing those lay-offs to technology is simply clever PR, and helps boost the share price.
So why does the faith in AI remain so strong?
The dubious hype doesn’t help. Every few weeks a new AI model appears, and smashes industry benchmarks. xAI’s Grok 4 did just that last week. But these are deceptive and simply provide more confirmation bias.
“Every single one of them has been wide of that mark. And not one has resolved hallucinations, alignment issues or boneheaded errors,” says Marcus.
Not only is generative AI unreliable, but it can’t reason, as a recent demonstration showed: OpenAI’s latest ChatGPT4o model was beaten by an 8-bit Atari home games console made in 1977.

“Reality is the ultimate benchmark for AI,” explained Chomba Bupe, a Zambian AI developer, last week. “You not going to declare that you have built intelligence by beating toy benchmarks … What’s the point of getting say 90pc on some physics benchmarks yet be unable to do any real physics?” he asked.
Then there are thousands of what I call “wowslop” accounts – social media feeds that declare amazement at breakthroughs. As well as the vendors, a lot of shadowy influence money is being spent on maintaining the hype.
This is not to say there aren’t uses for generative AI: Anthropic has hit $4bn (£3bn) in annual revenue. For some niches, like language translation and prototyping, it’s here to stay. Before it went mad last week, X’s Grok was great at adding valuable context.
But even if AI “discovers” new materials or medicines tomorrow, that won’t compensate for the trillion dollars that Goldman Sachs estimates business has already wasted on this generation of dud AI.
That’s capital that could have been invested far more usefully. Rather than an engine of progress, poor AI could be the opposite.
METR added an amusing footnote to their study. The researchers used one other control group in its productivity experiment, and this group made the worst, over-optimistic estimates of all. They were economists.
 
I’m amazed at the reality gap between what I e found generative ai useful for and what my senior management thinks it’s useful for. I am encouraged to ‘use it on every aspect!’ Of my job buy it’s just only useful for a very few things.
I keep asking for examples of how people are using it and get none. It’s great for a very few aspects and useless for day to day tasks
Maybe they're being 'encouraged' by their higher ups to force people to use it for everything, not because it's great, but because it cost a lot of money and they need to feel like they are getting use out of their 'investment.'

From what I've read on AI, a lot of companies bought into the hype and I'm assuming they don't want to feel like they 'wasted' their (investors') money.
 
Maybe they're being 'encouraged' by their higher ups to force people to use it for everything, not because it's great, but because it cost a lot of money and they need to feel like they are getting use out of their 'investment.'

From what I've read on AI, a lot of companies bought into the hype and I'm assuming they don't want to feel like they 'wasted' their (investors') money.
That's an aspect to it, but there's also the element where the higher ups assume "making the staff use AI" = "training the AI that will replace them". When unsurprisingly, people getting forced to use AI end up spending more of their time wrangling the AI or manually fixing the outputs, they get frustrated and think it's just an issue with people being bad at using AI or being luddites.

The higher ups do generally seem to think we're a few years away from Star Trek, where they can go "Computer, prepare a quarterly financial report and identify six areas for efficiencies, then renegotiate xyz supplier contracts, source an additional premises for expansion that meets all our needs and then find and onboard 20 more clients with a tailored sales pitch" and it'll do it with no further human involvement. LLMs just aren't capable of that, although perhaps agentic AI may start giving the impression of being capable of that (up till the moment it starts telling suppliers it needs 500 tungsten cubes to fulfill an order with an imaginary client).
 
“Recognising that it sucks in your own speciality, but imagining that it is somehow fabulous in domains you are less familiar with”, is something he calls “ChatGPT blindness”.
Interesting 🤔 that journos are an exception to the former and haven't stopped whining about the nu gr8 replacement.
 
  • Like
Reactions: LilShotaBoy
Machine learning tech is in a bubble just like PCs in the latgloomy.
Machine learning tech is not at all in a bubble. Large Language Models are one tiny subset of machine learning and they receive an outsized portion of the attention and money.
and whatever is left is basically only going to be used to make life more hellish for the common man, more propagandized, and more psyop'd by the niggers running things like machine learning search engines.
the rest of this post is just schizo nonsense. Your doomer programming is 5 years out of date; woke is dead and the evidence is right in your face. Google search is permafucked and has been for years. It no longer works and Google is struggling as a company. They are doomed to the path of IBM. LLMs are getting traction because are a better search engine than the ones that currently exist.

They also reflect the training data fed to them. If the internet has been censored to allow only left-wing viewpoints, it's going to reflect that in its answers. That doesn't at all mean it's permanently stuck on the same dataset, they retrain the models constantly.
Shit is tailored and controlled from top to bottom to only give the goyim the most Kosher of results, and no one will be able to get to the knowledge of the truth.
I absolutely despise posts like this. Just kill yourself already if everything is so doom and gloom.

What a pathetic existence you seem to live, seeing kikes around every corner, controlling every aspect of your life. It is mental illness. Newsflash, Hitler literally gathered the jews of Europe up, loaded them onto cars like cattle and sent them straight to death camps and they followed blindly by the millions, like livestock. They're not controlling the world.

The browbeating of White people in America, the racial replacement of us, the destabilizing of the country all benefit communists just as much as jews. Jews are just a convenient scapegoat because they're paranoid as fuck, can be easily provoked into sperging out and circle the wagons over anything and everything.

Don't let the jews live rent free in your head, you know if the shoe were on the other foot they would charge you rent.
 
Maybe they're being 'encouraged' by their higher ups to force people to use it for everything, not because it's great, but because it cost a lot of money and they need to feel like they are getting use out of their 'investment.'

From what I've read on AI, a lot of companies bought into the hype and I'm assuming they don't want to feel like they 'wasted' their (investors') money.
Seems to be right. A lot of companies have the urge to use ROI and accounting methods as a way to view efficiency in a production setting. They buy machines or software and force the workers to utilize it in order to make the investment seem more profitable. This mostly stems from the fact that they have to somehow validate their investments to higher management; even at the cost of overal productivity.
Also it is the IT-problem in manufacturing where management gets misled by buzzwords into thinking that this new software will make them earn more profits than before. Even the basic implementation of MRP-systems for manufacturing never correlated with increased efficiency and profits.
 
There was massive over hiring during covid. My theory is that all the mass layoffs lately are from that, not AI "taking jobs". It does give a very convenient smokescreen to cull all the retards though, saying they have found "new AI efficiencies" instead of the reality which is cutting overpaid dead weight. If they're still requesting H1Bs, AI is not taking the jobs.
 
AI can be handy to troubleshoot issues with things like coding or data manipulation. The issue is that you need to already know your inputs, your desired outputs, and the basics of the software, to get any use from AI.

But this isn't what it's been marketed as. We've been told for two years that AI will replace half the job market, revolutionize science, destroy all writers and artists, usher in a new golden age, and fuck your wife. Because of this potential, governments need to double their electric power generation and start building data centers everywhere. Now the stock market is being carried by Nvidia.
 
AI and LLMs can actually have real useful uses especially for things like analyzing data on a large scale to find patterns. There is even a guy who made an AI to play Smash Bros Melee called Philip.ai and it plays at high enough level to make top players sweat. Cool applications like that are not being invested in because corporations are retarded and all they want is to replace their Indians with AI.

The scariest part of AI is all the NPCs who use it to think for them. Frank Herbert predicted this all the way back in the 60s.

images (19).webp
 
Interesting 🤔 that journos are an exception to the former and haven't stopped whining about the nu gr8 replacement.

The journos were already being wiped out before AI and AI is just accelerating the process. Honestly nearly forever a tremendous number of jobs in "journalism" involved taking an existing article from one of three sources (AP, Reuters, New York Times) and rewriting it to (a) hide that it was swiped and (b) insert additional bias into it.

What AI is also doing is eliminating a whole set of low-paid writing jobs in journalism where people were being paid tiny money to create online filler content to back up a clickbait headline.
 
I’m amazed at the reality gap between what I e found generative ai useful for and what my senior management thinks it’s useful for. I am encouraged to ‘use it on every aspect!’ Of my job buy it’s just only useful for a very few things.
I keep asking for examples of how people are using it and get none. It’s great for a very few aspects and useless for day to day tasks

You have to ask normies. A huge number of them already use it for everything. You think it's useless for day to day tasks because you operate on a different standard than them.

As one example: The percentage of the population that can tell the difference between actual decent writing and "AI slop" writing is actually tiny. We just don't realize this, because we surround ourselves with other people who have similar literacy levels. We all complain about the stupid em-dashes, normies using AI are just happy that their email/text message/eulogy makes basic sense and doesn't have any glaring grammar mistakes.
 
You think it's useless for day to day tasks because you operate on a different standard than them.
This specifically cannot be overstated enough. If you're smart and capable, asking AI to help schedule your day will look retarded - You know what you're doing, and you are capable of thinking at least three steps ahead and sequencing things appropriately. To you, the AI reminding you to write a grocery list before going to the store is a no-shit moment - We've all forgotten one or two items before, and learned to avoid it. You probably sit down with the flyers, physical or online, and make a quick plan based on sales and preferences of what to buy so you can make enough dishes for the week, plus incidentals you might need.

To someone who's too stupid to do that and only goes to the store after realizing their fridge is completely empty? The AI prompting them to make a shopping list, then explaining to do the above when they ask "How do I make one" is absolutely genius tier shit.

AI isn't particularly smart. But the average person is particularly retarded and listless, whereas AI will never ghost your messages or judge you. The smoothbrains really do love it.

There was massive over hiring during covid. My theory is that all the mass layoffs lately are from that, not AI "taking jobs". It does give a very convenient smokescreen to cull all the retards though, saying they have found "new AI efficiencies" instead of the reality which is cutting overpaid dead weight. If they're still requesting H1Bs, AI is not taking the jobs.
There's a couple pieces all moving in the AI craze, this is one of them - Overhiring was a real thing, and it came on the backs of overinvestment in general as well. Most of that covid stimulus ended up in the stock market, soaked right back up by the companies. When AI blew up, we got a few different strains of companies all taking different approaches to it. Everyone wanted to justify getting a bigger share of that investor money.

For companies providing the software itself (OpenAI, etc), that was easy - sell AI on saving money. Even if its slim potentials, like a 1% chance that LLM models lead to a true AGI that can replace people, that's a one percent chance to have a multi-million percent return on investment if it works. Not a hard pitch to sell. And if the AI market dies off, their minimal costs are incredibly low, they can pare back staff, development, etc, and just focus on licensing the finished models. If they can't afford to make new ones, neither can anyone else, so they're still sole kings of the space. A depressing, shell of the former glory like IBM or Blackberry, but still functional. They'd return back to smaller, elite research institutions playing around with the next potential avenue to develop AGI, and we'd probably hear from them in another decade.

For the companies providing infrastructure and services via the software (Microsoft, Amazon, etc), it was a super simple reason to justify massive datacenter expansions and operations. If that 1% chance struck true, again, multi-million percent return on investment, that computing hardware would be worth absolute gold. And unlike the AI software companies, they also have a really promising fallback plan. Even if that 1% doesn't manifest, and they have all these fucking datacenters, datacenters are still their absolute #1 source of revenue. Azure and AWS are literal money printers, and having more of it is never a bad thing. They'd be doing so at unfavorable rates if it turns out they've overinvested, but that's not so bad. Neither of these companies are liable to disappear anytime soon, so investors have a slim chance at literally insane returns in the short term, and very good chance of decent returns on the 20 year outlook. Not a hard investment to add to the portfolio for most institutions.

For companies using the software and services, it gets a bit more interesting. The appeal of replacing staff is always there - not just for immediate staff reduction, but removal of the costs of turnover as well. Employees are expensive both to keep and replace, between situational training, knowledge transfer, severance, headhunters and talent firms, it all adds up, but AI never quits. What mattered more was the optics of it. No company was safe from the overinvestment craze off the back of covid, and the inevitable market correction was looming on them all. The best way to keep those inflated stock prices was to justify how you were, in fact, actually worth that big investment and worth continuing it. Easiest way to do that was to latch onto the promising longshot of AI, so you look like your part of a potential industry leading trend. Its far cheaper for a company to allocate tens of millions of capital into some performative AI standup than it is to risk losing hundreds of millions in potential fundraising opportunity from future stock sales. So they bought in these services and plans, to look like they were ahead. And if they sunk that money, they might as well make the most of what they got, hence a lot of forced adoption. Investors want to see that the money is doing things, so being able to say 70% of our staff use AI in their work, even if the AI is just an AI search bar somewhere on an internal tool, it looks good to them. And if the AI stuff crashes and the market corrects, then all it really costs them is the correction that was inevitable otherwise. Its not that there's no risk here, just that this is the lowest risk path, everything else is guaranteed outcomes at least as bad as this one, but with much lower chances of good results.

Amusingly, the company shouldering the most risk here is Nvidia and the other AI hardware companies. The equipment is extremely specialized, and the RnD is extremely expensive. If the floor falls out of the AI market, they're not in the position of the consuming market to continue their business model without AI, they're not in the position of the infrastructure companies where they can just cut back spend and ride the compute they bought in advance, and they're not in the place of the software companies who can just cut to the bone and coast on the old works existence. Manufacturing has huge overhead, significant minimums, and constantly needs to advance. If the AI space becomes dead air and their ASIC's go back to being mostly scientific resources and CG Renderfarm hardware, they're left holding a hundreds of billions of dollars bag. Their existing markets are mostly saturated in the gaming and non-AI datacenter spaces, this was their sole growth realm. They'll crash hard if the AI market goes dead.

And that is my expectation, it will go dead. The evidence is already clear that LLM models cannot reach or mimic AGI level, LLM's are useful for limited cases but a dead end in the chase for true intelligence. The only question is how far will those limited use cases be stretched, and for how long.
 
revolutionize science,
They won’t. Quantum type computing will if it ever happens. To be able to construct a mammalian system in silico and see what happens when a drug is introduced would be incredible. That would require a complete napping of every molecular interaction. I’m not even sure it’s possible but if it is it will need mind boggling computing power .
like analyzing data on a large scale to find patterns.
Now they ARE good for that. In silico drug design, combing through molecule libraries, analysing array data - there’s a utility there, along with image analysis, pathology as well. I saw an article where it’d been used to find single sperm in a sample from a guy with almost total lack of sperm. But all these are brute forcing ‘throw enough time at it’ problems. Not genuine novel thought
You think it's useless for day to day tasks because you operate on a different standard than them.
I hadnt even considered that. This is depressing as. The level it operates on is low. Are management that low performing they think it’s a boon??
We keep being told ‘it can write your emails!’ It. A streamline your day! As if an email is a challenge. Even worse I can see that exec level IS using it to write their Emails because they all sound like the same ai slop voice now. it’s use of language is dogshit
You can even see it change over time in rhe slop press. Look at the Dm. You can track by the keywords it uses. ‘Beloved’ was in vogue last month and now it’s ‘disgusting.’ There was a trend a few months ago for ‘x word.’ So ‘mum’s four word reply to…’ or ‘popstar’s five word SnapBack…’ it’s obvious something is churned out ai slop.
There are places where machine learning works really really well. Day to day office tasks is not one of them. One of our systems which has been unveiled to great fanfare barely works better than the spreadsheet it replaces. At least I could see where the data came from in the spreadsheet. The system has a bot you can ask and we are told it’s all powerful and can answer questions. I’ve tested it. I’ve asked it very simple things. What happens if this is over resourced by x hours a month? Nonsense answer. I’ve tried simpler things. It cannot answer them. It’s useless.
I tried asking the google one to convert between kg and stone and pounds and it was wrong.
 
We keep being told ‘it can write your emails!’ It. A streamline your day! As if an email is a challenge. Even worse I can see that exec level IS using it to write their Emails because they all sound like the same ai slop voice now. it’s use of language is dogshit
True but look at most corporate emails. It's a few sentences of content that takes 10 pages because longer emails are more impressive at their level. So get AI to pad out your paper. Then everyone who reads it uses the AI summary because who has time to read 10 pages of slop.
 
Do you have a larger version of that? I had a quick google but can't find one.

Now they ARE good for that. In silico drug design, combing through molecule libraries, analysing array data - there’s a utility there, along with image analysis, pathology as well. I saw an article where it’d been used to find single sperm in a sample from a guy with almost total lack of sperm. But all these are brute forcing ‘throw enough time at it’ problems. Not genuine novel thought

I was listening to something the other day that someone said it's going to be great for Telescopic hunting of Asteroids or other space objects, we've apparently got the imaging capability cheap now and automated in pointing teams of smaller optical telescopes at the sky, the harder an expensive part is image analysis like comparing photographs of the same part of the sky from 50 nights takes time an effort an small details are often missed but LLM's can play spot the difference really well over data from larger sets - they are just trying to work out the methodology now to make sure it doesn't confuse things or flag too many false positives.
 
Do you have a larger version of that? I had a quick google but can't find one.

Sorry I just clicked on Google Image which is an absolute pain in the fucking ass because it literally never gives the full image or makes it easy to find the original through even a link. Total Pajeet death cannot happen soon enough.
 
I hadnt even considered that. This is depressing as. The level it operates on is low. Are management that low performing they think it’s a boon??

I've worked for the largest corporations on the planet. Most of the managerial class is completely surperfluous and the memes about bullshit e-mail jobs and jew daycare are 100% accurate,
 
I've worked for the largest corporations on the planet. Most of the managerial class is completely surperfluous and the memes about bullshit e-mail jobs and jew daycare are 100% accurate,

The managerial paralysis of the large corporations is why almost anything innovative has to be done in smaller start-up companies. The managers and the managerial class are bad, but its also laws, rules and the threat of lawsuits that just makes it impossible to do much of anything.
 
Back