Skip to main content

How Much Is That AI Subscription In My Windows?

Inference might be agreed to be the new battleground where the next AI winners emerge, but even before we reach that brave new frontier, the fight is on at the intelligence cash register. The fight for licensing and subscription deals continues apact this month.

Anthropic’s popular Claude Code has been at the forefront of many issues in this space with developers reportedly using up their entire month’s usage limits in days, sometimes even hours.

Single Prompt Sinkhole

Claude Code user Pristine_Ad2701 posted on Reddit to explain that their $100 plan was almost swallowed up by a single simple code requisition prompt.

“One prompt and it uses 37% [of my] 5-hour limit, after writing literally NORMAL things, nothing complex, literally, CRUD operations, switching to sonnet, it was currently on 70%,” they wrote.

None of which really tallies with what the company is stating in official terms. According to an SSD Nodes blog, “According to Anthropic’s own data, the average Claude Code user costs about $6 per developer per day, with 90% of users staying under $12/day. At full-time usage with Sonnet 4.6, that projects to roughly $100–$200 per developer per month, which is exactly where the Max plan sits.”

This is said to be the “break-even point” where subscription pricing starts outperforming direct API billing for heavy individual users.

Toen Tariff Thresholds

Meanwhile, over at OpenAI, the waters aren’t exactly calm, but the organization is looking to address token tariff tribulations with its own go-to-market approach. OpenAI’s Pro ChatGPT tier arrived this week with an increased limit for Codex. 

As all good software engineers will know, OpenAI Codex is an artificial intelligence model that translates natural language instructions into functional code.

A Bargain For ‘High-Effort’ Users

Posting on X, OpenAI has said that its is now updating ChatGPT Pro and Plus subscriptions to support the growing use of Codex. The company is introducing a new $100/month Pro tier. This new tier is said to offer five times more Codex usage than Plus and is best for what the company defines as “longer, high-effort” Codex sessions. 

In ChatGPT, the new Pro tier still offers access to all Pro features, including the exclusive Pro model and unlimited access to Instant and Thinking models.

“To celebrate the launch, we’re increasing Codex usage for a limited time through May 31st [2026] so that Pro $100 subscribers get up to 10x usage of ChatGPT Plus on Codex to build your most ambitious ideas,” stated OpenAI, on the social network formerly known as Twitter.

Ask an AI, About AI

With so many factors, vector and actor involved here, it’s tough to know where the dust will settle in this marketplace. It’s almost tempting enough to ask ChatGPT something like – why is there so much confusion in the software developer marketplace around pricing and usage limits for coding assistant technologies?

Spoiler alert, we did. 

The service thinks that pricing models “don’t match how developers think” and that the cost structure across the industry is both volatile and opaque, i.e., different tools optimize for different use cases. We can also remind ourselves that “unlimited” rarely means unlimited, multiple dimensions are often bundled into single plans, and that there is misalignment and confusion fuelled by the fact that there is no standard terminology across vendors.

Anchored Down at Average

For a more erudite (and rather more human-value) answer, Mitch Ashley, VP and practice lead for software lifecycle engineering at The Futurum Group, helps lay down the real state of the nation.

“Subscription pricing for AI coding tools is being stress-tested by the workloads it was built to serve,” explained Ashley. “The gap between vendor-stated consumption averages and actual power-user experience reflects how poorly linear pricing maps to agentic, context-heavy development sessions.

“Vendors that anchor pricing to average consumption will lose their highest-value users to API billing or competitors with more flexible tiers. The developer burning through a $100 plan in a single session is the retention risk that determines who wins the subscription battle.”

All of which shenanigans are happening at the same time as the industry puts a renewed focus on AI agent observability in order for us to understand what services are doing when we do use up our precious subscription limits on them.

I Can Has Cheezburger?

As reported here on DevOps.com, a survey of 628 enterprise IT leaders conducted by The Futurum Group finds well over a third (36%) plan on spending more than $1 million on observability in 2026, with 7% planning to spend in excess of $5 million.

There’s a price to pay for AI, and we’re some way off being able to work with a single clear à la carte menu. I’ll just have the cheeseburger for now, then, please.



from DevOps.com https://ift.tt/zPeUxl2

Comments

Popular posts from this blog

Claude Code’s Ultraplan Bridges the Gap Between Planning and Execution

Planning a complex code change is hard enough. Reviewing it in a terminal window shouldn’t make it harder. Anthropic is addressing that friction with a new capability called Ultraplan, currently in research preview as part of Claude Code. The feature moves the planning phase of a coding task from your local terminal to the cloud — and gives developers a richer environment to review, revise, and approve a plan before a single line of code changes. It’s a small workflow shift with real practical value, especially for teams working on large-scale migrations, service refactoring, or anything that requires careful coordination before execution begins. How it Works Ultraplan connects Claude Code’s command-line interface (CLI) to a cloud-based session running in plan mode. When a developer triggers it — either by running /ultraplan followed by a prompt, typing the word “ultraplan” anywhere in a standard prompt, or choosing to refine an existing local plan in the cloud — Claude picks u...

Java 26 Arrives With AI Integration and a New Ecosystem Portfolio — What It Means for DevOps Teams

Oracle released Java 26 on March 17, 2026, and while every six-month release comes with its own set of improvements, this one carries a broader message: Java isn’t just keeping pace with the AI era — it’s actively positioning itself as the infrastructure layer where AI workloads will run. For DevOps teams managing large Java estates, that’s worth paying attention to. The Scale of What You’re Already Running Before getting into what’s new, it helps to remember what’s already in place. According to a 2025 VDC study, Java is the number one language for overall enterprise use and for cloud-native deployments. There are 73 billion active JVMs running today, with 51 billion of those in the cloud. That scale matters when you’re thinking about where AI fits in. Most of the systems where agentic AI will eventually operate — transactional platforms, backend services, data pipelines — are already running on Java. The question for DevOps teams isn’t whether to adopt Java for AI. It’s how to ...

Gremlin Adds Detected Risk Tool to Chaos Engineering Service

Gremlin's risk detection capability in its chaos engineering service automatically identifies issues that could cause outages along with recommendations to resolve them. from DevOps.com https://ift.tt/iaw9Q7D