We’re hitting a pivotal moment in artificial intelligence right now. The latest financial disclosures from the front lines of the AI arms race — OpenAI Group PBC and Anthropic PBC — don’t just give us a peek under the hood; they expose the core tension shaping the entire industry. The smartest machines ever built are also the most expensive to create and operate in the history of computing.
That model is breaking. AI isn’t pure software; it’s infrastructure-as-a-service with a heavy dose of industrial-era physics. It is compute-bound, energy-intensive and capital-heavy. When OpenAI projects a staggering $121 billion in compute spend by 2028, it isn’t an outlier. It’s a signal that we have entered the era of “Infrastructure Gravity.”
Both OpenAI and Anthropic are now reporting profitability in two distinct ways: with training costs excluded, and with training costs included.
That alone should stop you in your tracks. It reveals that training costs are no longer a “R&D expense” — they are the cost of goods sold. Strip them out, and you have something that looks like software as a service. Put them back in, and you’re staring at one of the most capital-intensive industries ever created.
The 2025-2026 financial reality:
Annualized revenue: OpenAI ~$13 Billion | Anthropic ~$7 Billion
Valuation: OpenAI ~$300 Billion | Anthropic ~$183 Billion
Compute/training intensity: OpenAI – High (Consumer + Enterprise) | Anthropic – High (Enterprise + API)
By 2026, inference is projected to account for 65% of all AI compute, and over a model’s lifetime, it can represent 80% to 90% of total costs. We’ve seen this with GPT-4: a ~$150 million training cost ballooned into an estimated $2.3 billion in inference costs within two years. This is a 15x multiplier that software-centric VCs aren’t used to seeing.
We are seeing a clear divergence in how these giants chase the “means of production”:
- Anthropic: Leaning heavily into the enterprise and API layers. Approximately 80% of its revenue now comes from business customers and API calls (notably powering tools like GitHub Copilot and Cursor). This brings predictable revenue and higher margins.
- OpenAI: Straddling both worlds. It’s building a consumer platform (ChatGPT has more tan 800 million weekly active users) while maintaining a massive enterprise footprint. For them, the free tier isn’t a gift; it’s a data-collection infrastructure.
If you zoom out, this starts to look familiar. We’ve seen this with Amazon Web Services Inc., Microsoft Azure and the telecom buildouts of the ’90s. Massive upfront capital, long payoff cycles and inevitable market consolidation.
The difference? Speed. AI is compressing decades of infrastructure buildout into a few years. This is why the “Big Four” hyperscalers are expected to spend more than $650 billion into capital expenditures by 2026 alone.
The real battle: Bending the cost curve
The game ends in one of two ways. If hardware efficiency (such as Google LLC’s TPUs or Nvidia’s Blackwell) can drop the cost of intelligence faster than demand rises, the model becomes unstoppable. If not, we are looking at massive margin compression and a brutal era of consolidation.
The bottom line
Don’t get distracted by the demos. In this new era, the companies that win won’t just have the best models. They’ll have the best economics. Focus on the stack: compute, cost, distribution, monetization. That’s where the $100 billion question will be answered.
Image: News
Support our mission to keep content open and free by engaging with theCUBE community. Join theCUBE’s Alumni Trust Network, where technology leaders connect, share intelligence and create opportunities.
- 15M+ viewers of theCUBE videos, powering conversations across AI, cloud, cybersecurity and more
- 11.4k+ theCUBE alumni — Connect with more than 11,400 tech and business leaders shaping the future through a unique trusted-based network.
About News Media
Founded by tech visionaries John Furrier and Dave Vellante, News Media has built a dynamic ecosystem of industry-leading digital media brands that reach 15+ million elite tech professionals. Our new proprietary theCUBE AI Video Cloud is breaking ground in audience interaction, leveraging theCUBEai.com neural network to help technology companies make data-driven decisions and stay at the forefront of industry conversations.
