OpenAI and Anthropic hike API costs to boost profitability ahead of anticipated public offerings
OpenAI and Anthropic pivot toward aggressive monetization, ending the subsidized AI era to secure profitability ahead of public offerings.
May 10, 2026

The era of subsidized artificial intelligence appears to be drawing to a close as the industry's two most prominent players, OpenAI and Anthropic, pivot toward aggressive monetization strategies ahead of anticipated public offerings. While early phases of large language model development were defined by rapid price cuts and efficiency gains, the launch of GPT-5.5 has signaled a sharp reversal. Recent analysis of real-world usage data suggests that the latest flagship model from OpenAI carries a significantly higher financial burden than its predecessor, GPT-5.4, with effective costs rising between 49 and 92 percent depending on the complexity and length of the user's input.[1][2] This shift represents a fundamental change in the economic relationship between AI providers and the developers who build upon their infrastructure, marking the end of the "growth at all costs" phase and the beginning of a high-margin, enterprise-focused era.
When OpenAI officially unveiled GPT-5.5 last month, the company acknowledged a doubling of the list price, setting input tokens at five dollars per million and output tokens at thirty dollars per million.[1] To soften the blow to developer budgets, OpenAI representatives claimed that the model’s improved reasoning capabilities and reduced verbosity would naturally result in shorter, more concise responses that would partially offset the price hike. However, an exhaustive analysis of real usage data conducted by OpenRouter tells a far more complicated story. By tracking a switcher cohort of users who migrated their primary workflows from GPT-5.4 to GPT-5.5, researchers found that while the model is indeed more efficient in specific high-context scenarios, it is frequently more expensive in common, mid-range tasks. The data indicates that for prompts exceeding 10,000 tokens, GPT-5.5 produces completions that are 19 to 34 percent shorter, which provides some relief for large-scale document analysis. Conversely, for prompts in the 2,000 to 10,000 token range, completions were found to be 52 percent longer than the previous version, leading to the upper end of the 92 percent cost increase.[1][3]
The discrepancy between list price and actual billed cost highlights the nuance of modern tokenomics. For short-form inputs under 2,000 tokens, where response length remained largely static between versions, the effective cost nearly doubled overnight.[1] This has hit small-scale developers and consumer-facing applications the hardest, as these workflows rarely benefit from the long-context efficiencies OpenAI highlighted during the launch. The industry is witnessing the emergence of a bifurcated market where the most capable frontier models are being priced out of reach for general-purpose applications, effectively reserved for high-value professional workloads such as agentic coding, legal discovery, and specialized scientific research. This pricing volatility is also catching many organizations off guard, with recent reports indicating that 80 percent of companies are missing their AI infrastructure cost forecasts by more than 25 percent, leading to a direct erosion of gross margins for AI-native startups.
OpenAI is not alone in its pursuit of higher revenue per token. Anthropic recently introduced Claude Opus 4.7, which at first glance appeared to maintain the sticker price of its predecessor at five dollars per million input tokens and twenty-five dollars per million output tokens.[4][5] However, developers quickly identified what has been termed a tokenizer trap. The new version of Opus utilizes a redesigned tokenizer that produces up to 35 percent more tokens for the identical block of raw text compared to previous versions.[5][4] This architectural change acts as a stealth price hike, ensuring that even if the rate card remains the same, the monthly bill for an enterprise customer will inevitably rise. Furthermore, Anthropic’s decision to shift its Claude Code environment to a higher-intensity reasoning mode by default has further inflated token consumption for software engineering tasks.[5] These tactics suggest a coordinated industry move toward enhancing bottom-line figures as both companies prepare for some of the most highly anticipated initial public offerings in the history of the technology sector.
The financial pressure driving these decisions is immense. Anthropic is reportedly operating at an annualized revenue run rate of over 30 billion dollars, a staggering jump from the nine billion reported just a year prior.[6][7] Yet, despite this growth, the cost of maintaining and scaling the specialized GPU clusters required for these frontier models continues to weigh on profitability. With valuations for these firms now reaching toward the one-trillion-dollar mark on secondary markets, the need to demonstrate a clear path to self-sustaining profitability is paramount. For investors, the price hikes are a sign of market maturity and pricing power; for the ecosystem of developers, they are a source of increasing anxiety. The risk of vendor lock-in has never been higher, as migrating complex agentic workflows between different model families is a technically grueling and expensive process, often leaving companies with little choice but to absorb the rising costs.
In response to the escalating cost of flagship intelligence, a secondary market of budget-tier models has begun to flourish. Models such as GPT-5 Nano and Gemini 3.1 Flash-Lite have seen explosive adoption for high-volume, low-complexity tasks like classification and simple summarization. These models are often priced at a fraction of a cent per million tokens, creating a massive 500-fold price gap between the most affordable and most capable tiers of AI.[8] This has forced a new level of architectural sophistication upon developers, who must now implement complex routing logic to ensure that only the most difficult queries are sent to expensive models like GPT-5.5 or Opus 4.7. The era of using a single flagship model for every task is effectively over, replaced by a tiered approach where cost management is as critical a skill for an AI engineer as prompt engineering or fine-tuning.
The implications for the broader SaaS landscape are profound. As AI providers raise prices, the software companies that have integrated these models into their products are being forced to decide whether to raise their own subscription fees or absorb the costs and risk investor backlash. We are already seeing the impact of this in the public markets, where established software giants have seen their valuations waver as the high cost of AI integration begins to impact their financial reporting. The current trend suggests that the industry is moving away from the democratized access that characterized the early 2020s toward a more traditional enterprise software model where the highest levels of machine intelligence are a premium commodity. As OpenAI and Anthropic march toward their eventual market debuts, the priority has clearly shifted from winning the most users to extracting the most value from every token generated. Accuracy and reasoning power are no longer the only benchmarks that matter; for the modern enterprise, the most important metric is now the return on every dollar spent on the API.