The launch of OpenAI's long-awaited GPT-5 model last week marked an ignominious chapter for the company. While it boasted modest performance upgrades on paper, the actual experience of using it left many fans sorely disappointed.
Common criticisms were that its answers were too short and its writing was noticeably worse and devoid of personality. On top of that, many noted that despite its "PhD level" intelligence, it still made dumb errors like insisting there are three Bs in the word "blueberry."
It was such a letdown that fans demanded that OpenAI give them back access to the previous model GPT-4o, which the company boldly removed with the launch of GPT-5. And stunningly, OpenAI capitulated to fans' demands.
In sum, it did not feel like a "significant step along the path to AGI," as CEO Sam Altman bragged it was.
There may be an interesting reason behind why the ChatGPT upgrade feels so lackluster. As The Register speculates, it's likely because GPT-5 is really meant to be "less of an advancement and more of a way to save compute cost."
It's a striking theory, highlighting how the AI industry continues to burn through billions of dollars, eclipsing moderate revenues. OpenAI could be looking to pump the brakes to better its chances of turning a profit before the end of the decade, but whether that's even a possibility at this point remains a point of contention.
GPT-5 isn't a single model — it's actually a tag team of a lightweight model to handle basic requests, and a beefier one to tackle complex "reasoning" tasks. A separate router model chooses which of the two LLMs should tag in, based on the prompt.
As it turned out, the router model sucked at its job, and "broke" on launch day, Altman claimed, making GPT-5 seem "way dumber."
It's now back online with some purported improvements, but has done little to conciliate fans' fury at having their choices limited.
The important takeaway from this is that deploying this "autoswitcher" strays from OpenAI's approach in the past, The Register notes, which allowed paid users to simply select which model they wanted to use, instead of it being chosen automatically. Keeping all those models online, however, is expensive, reinforcing the idea that the ChatGPT maker is undergoing some belt-tightening by getting rid of them — though it has since reversed course and restored access to paid users.
There are other blatant signs of cost-cutting, like a severe limit of just ten messages per hour for free users. OpenAI is also keeping the model's "context window" — essentially its memory — the same as before, limited to 32,000 tokens for Plus users and 128,000 for Pro. Fans, especially those in the Plus tier — which at $20 is the only affordable tier for most, as Pro will set you back $200 per month — have been begging for an upgrade in this area.
GPT-5 representing behind-the-scenes pruning at the company certainly been the theory held by many of its fans, with one of the top posts on the r/ChatGPT subreddit averring that GPT-5 is "clearly a cost-saving exercise."
"They removed all their expensive, capable models and replace[d] them with an auto-router that defaults to cost optimisation," the user wrote. "That sounds bad, so they wrap it up as [GPT-5] and proclaim it's incredible."
"Feels like cost-saving, not like improvement," surmised another.
Unglamorous as it is, cost-cutting at this moment makes sense from OpenAI's point of view. It's facing more competition than ever and is under increasing pressure to find a way to turn its business model profitable. Its anticipated valuation of some $500 billion comes with the implicit expectation that it will figure out how to make money soon.
But the company clearly underestimated just how wildly attached fans would become to the quirks of its older models, even if they are nominally inferior — and that's not a problem that's going to go away anytime soon.
More on AI: Scientists Are Getting Seriously Worried That We've Already Hit Peak AI
Share This Article