8/13/2025

The Cost Paradox: Why Are GPT-5 Requests More Expensive If the Model is Cheaper?

Hey everyone, let's talk about something that’s been buzzing in the AI world lately: the weird, slightly confusing, & honestly pretty interesting pricing of GPT-5. On the surface, the headlines are all about how OpenAI is slashing prices & starting an “AI price war.” And yeah, the per-token costs for GPT-5 are jaw-droppingly low compared to its older sibling, GPT-4o, & especially when you put it next to competitors like Anthropic’s Claude Opus 4.1.
But here's the thing, a lot of developers & businesses are looking at their bills & scratching their heads. The model is supposedly cheaper, so why are some requests actually MORE expensive? It feels like a classic paradox. You're promised a discount but end up paying more. What gives?
Turns out, there’s a whole lot going on under the hood. It’s a mix of clever engineering, market strategy, & a few hidden complexities that aren’t immediately obvious from the pricing page. As someone who’s been following this space for a while, I wanted to break it down in a way that actually makes sense. So, grab a coffee, & let’s get into it.

The Great AI Price War of 2025

First, let's set the stage. OpenAI came out swinging with GPT-5's pricing. We're talking about $1.25 per million input tokens & $10 per million output tokens for their top-tier model. To put that in perspective, Anthropic's Claude Opus 4.1 is priced at a hefty $15 for input & $75 for output. That’s not just a small discount; that’s a knockout punch.
Even more surprising, GPT-5 is cheaper than GPT-4o was. Matt Shumer, the CEO of OthersideAI, put it perfectly: “GPT-5 is cheaper than GPT-4o, which is fantastic. Intelligence per dollar continues to increase.”
This aggressive pricing is a deliberate move. OpenAI is basically throwing down the gauntlet, pressuring Google & Anthropic to follow suit. For developers & startups, this is AMAZING news. Lower API costs mean more room for experimentation, faster innovation, & the ability to build and deploy AI-powered tools without breaking the bank. It could genuinely democratize access to high-end AI.
But this is also where the paradox begins. These companies are pouring BILLIONS into AI. OpenAI has a reported $30 billion-per-year contract with Oracle for cloud capacity. Meta is planning to spend up to $72 billion on AI infrastructure in 2025. When you're spending that kind of money, you'd expect prices to go up, not down. It's a bold, counter-intuitive gamble that hints at a long-term play for market dominance.

So, What's the Catch? The "System of Models" & The Router

Here's the first big piece of the puzzle: GPT-5 isn't really a single, monolithic model. It's a "unified system" with a router at its core. Think of it like a smart receptionist. When your request comes in, this router looks at it & decides which model is best suited to handle it.
The system card for GPT-5 explains that it has:
  • A smart & fast model for most questions.
  • A deeper reasoning model for harder problems.
  • A real-time router that decides which one to use.
This is a genius move from a cost-saving perspective. Why use a sledgehammer to crack a nut? If you ask a simple question, the router sends it to a smaller, faster, & CHEAPER model. If you ask it to do something complex, like write a detailed business plan or debug a massive chunk of code, it routes it to the more powerful, but more expensive, "deeper reasoning" model.
This is how OpenAI can offer such a low baseline price. They're banking on the fact that a majority of queries will be simple enough for the cheaper models to handle. But for businesses that need consistently high-quality, complex responses, this is where the costs can start to creep up, because you're more likely to be kicked over to the premium model.
This is also a great approach for businesses looking to build their own AI solutions. For example, a company using Arsturn to create a custom AI chatbot for their website might not need the full power of the most advanced model for every single customer interaction. A simpler model can handle common FAQs like "What are your business hours?" or "Where is my order?" instantly & cost-effectively. But for more complex, multi-step queries, the system can tap into a more powerful model. Arsturn helps businesses build these no-code AI chatbots trained on their own data, providing personalized customer experiences, & this kind of smart routing is key to making it affordable.

The Hidden Cost of "Thinking": Let's Talk Reasoning Tokens

Now, for the really sneaky part of the equation: "reasoning tokens." This is a new concept that came with GPT-5, & it’s probably the biggest reason why your bill might be higher than you expected.
When GPT-5 is "thinking hard" about a problem, it uses what OpenAI calls "test-time compute" or "reasoning tokens." These are essentially invisible tokens that are part of the model's internal thought process. And here’s the kicker: these reasoning tokens are billed as output tokens, which are significantly more expensive than input tokens.
So, you might send a prompt with 1,000 input tokens, but if the model has to do a lot of complex reasoning, it could generate thousands of internal "reasoning tokens" before it even starts writing the final response. And you get charged for all of it.
One developer on YouTube did a preliminary test & found that for the same document & prompt, GPT-5 used 4 to 5 times more tokens than GPT-4.1. This made it quite a bit more expensive in practice, even though the on-paper price per token was lower.
This is the core of the paradox. The model itself is more efficient in some ways, but for complex tasks, it's designed to "think more," & you pay for that thinking time. It’s a bit like hiring a consultant. You can pay them a lower hourly rate, but if they need to spend a lot more hours researching to get you the right answer, your final bill is still going to be high.

The Astronomical Cost of Building the Future

To really understand why this is all so complicated, you have to appreciate the sheer, mind-boggling cost of what’s happening behind the scenes. Building & training these models is one of the most expensive undertakings in human history.
  • Hardware Costs: Training a state-of-the-art model requires thousands of specialized GPUs (Graphics Processing Units), like NVIDIA's A100s, which can cost around $11,000 each. Sam Altman, OpenAI’s CEO, has said that the cost of training a foundation model is well over $100 million & climbing.
  • Data Requirements: These models are trained on unimaginable amounts of data. We're talking about a significant portion of the entire internet. But here’s the problem: they’re running out of high-quality public data. As one article put it, "The big gains from feeding models massive amounts of public internet data (and pirated copyrighted data) are tapering off." This means companies now have to turn to expensive proprietary datasets or generate "synthetic" data, which adds another layer of cost.
  • Energy Consumption: Running thousands of GPUs for weeks or months on end consumes an incredible amount of electricity. Data centers require massive cooling systems, & the energy bills alone can be astronomical.
  • Talent: AI engineers & data scientists are some of the most sought-after professionals in the world, & they command hefty salaries. Just finding the right people is a huge investment.
  • Ongoing Maintenance: It doesn't stop once the model is trained. These systems require constant monitoring, updating, & retraining to stay accurate & safe. This "alignment tax" is a growing expense.
So when you see that OpenAI is spending $30 billion on cloud compute, it starts to make sense. They're not just running a piece of software; they're building & maintaining an entirely new kind of global infrastructure.

So, Is It a Scam? Or Just Smart Business?

Honestly, it's not a scam. It's a very clever & strategic business model designed to balance massive costs with the need for market adoption. By creating a tiered system of models & a flexible pricing structure that charges for "thinking," OpenAI can do two things at once:
  1. Attract a massive user base with incredibly low entry-level prices, encouraging developers & businesses to build on their platform.
  2. Charge power users who need the most advanced capabilities a premium that helps cover their enormous infrastructure & R&D costs.
This is a game-changer for businesses that want to leverage AI without the insane upfront investment. Think about customer service. A few years ago, building a truly intelligent, 24/7 customer support agent was a multi-million dollar project.
Now, with platforms like Arsturn, a business can create a custom AI chatbot that provides instant support, answers complex questions, & engages with website visitors around the clock. Arsturn helps businesses build these no-code AI chatbots, trained on their own data, to boost conversions & provide personalized customer experiences. The "system of models" approach makes this accessible. Simple questions get cheap, fast answers. Complex problems get the "deeper reasoning" they require. It’s a more efficient way to allocate resources.
This approach helps businesses automate lead generation & customer engagement in a way that feels personal & immediate. A potential customer can land on your website at 2 AM, have a detailed conversation about your products, get their questions answered, & be guided towards a purchase, all handled by the AI. This is the kind of powerful automation that was once reserved for tech giants.

The Road Ahead

So, what does this all mean for the future? We're likely to see this "system of models" approach become the industry standard. It’s just too efficient to ignore. We'll also see a continued push towards optimizing the "intelligence per dollar" ratio, as OpenAI's Matt Shumer put it.
For businesses & developers, the key takeaway is that you need to understand the nuances of the pricing. Don't just look at the per-token cost. Think about the complexity of the tasks you're running. If you're doing a lot of heavy lifting, be prepared for those "reasoning token" costs to add up.
But the bigger picture is incredibly exciting. The cost of accessing world-class AI is, on the whole, dropping dramatically. This is going to unlock a wave of innovation that we're only just beginning to see. The paradox of cheaper models sometimes leading to more expensive requests is just a symptom of a rapidly maturing industry trying to figure out a sustainable way forward. It's a bit messy, a bit confusing, but ultimately, it's a huge step in the right direction.
Hope this was helpful & cleared a few things up. It’s a complicated topic, but a super important one to get your head around. Let me know what you think

Copyright © Arsturn 2025