Grant Stellmacher
← Back to Research
AI & Business Models2026-04-0710 min read

OpenAI Is Burning $14 Billion a Year. The AI Business Model Problem Nobody Wants to Talk About.

TL;DR: OpenAI is projected to lose $14 billion in 2026 against $12.7 billion in revenue, burning cash faster than it can raise it. The core problem is that inference costs scale with usage in ways that make unit economics deeply unfavorable at current pricing. Every company building on top of AI APIs needs to understand what happens when the subsidized pricing era ends — because it will end, and the adjustment will be painful.

The numbers came out of The Information last month and the reaction was mostly shrugs. OpenAI: $12.7 billion in projected 2026 revenue. $14 billion in projected 2026 losses. Net loss exceeding revenue by ten figures.

The shrugs bother me. This isn't a startup burning cash to acquire users in a winner-take-all market — the classic Amazon playbook that everyone has been taught to accept. This is a company with no clear path to unit economics that work, operating infrastructure that gets more expensive as it gets more popular, and a fundamental misalignment between what customers pay and what delivery actually costs.

The people shrugging have mostly decided that OpenAI is "too important to fail" and that Microsoft, or sovereign wealth funds, or some future Microsoft, will keep the lights on. That may be right as a short-term prediction. It's a terrible framework for businesses making long-term bets on AI infrastructure.

What $14 Billion in Losses Actually Means

Let's put the numbers in context. OpenAI raised $6.6 billion in October 2024 at a $157 billion valuation. It raised an additional $40 billion in April 2025. That's nearly $47 billion in fundraising over eighteen months — more than most public companies generate in a decade of operations.

And it's still not enough to cover the losses.

The burn rate math is straightforward and brutal. At $14 billion in annual losses, OpenAI consumes roughly $1.17 billion per month. At that rate, the $40 billion raised in April 2025 covers about 34 months of losses — assuming revenue doesn't grow faster than costs, which is the central assumption worth examining.

Revenue is growing. OpenAI crossed $1 billion in monthly revenue in late 2024. The $12.7 billion 2026 projection implies substantial continued growth. The problem is that costs appear to be growing faster.

Why? Because the cost structure of large language model inference is fundamentally different from almost every other software business in history.

The Inference Cost Problem

Traditional software has near-zero marginal cost at scale. Once you've written the code and built the infrastructure, serving the ten-millionth customer costs essentially nothing more than serving the first. This is why software margins are extraordinary and why SaaS businesses trade at revenue multiples that seem absurd by any traditional metric.

LLM inference doesn't work this way.

Every query to GPT-4o or Claude 3.5 or Gemini Ultra requires actual computation — matrix multiplications running on expensive GPU hardware, consuming power, generating heat, requiring cooling. The marginal cost of serving a query is not zero. It's not even small. And crucially, it scales with the length and complexity of the query in ways that make pricing extremely difficult.

The current generation of AI models requires roughly 1-3 watts per token generated on modern hardware, depending on model size and batch efficiency. At scale, that adds up. OpenAI's compute costs alone — before salaries, facilities, bandwidth, or any other expense — are estimated at several billion dollars annually. The hardware depreciation on the H100 clusters needed to serve current demand is itself a nine-figure annual expense.

The pricing OpenAI charges customers doesn't come close to covering these costs at current volume levels. OpenAI charges roughly $15 per million output tokens for GPT-4o as of early 2026. Independent analysis suggests the fully-loaded cost to serve a million tokens — accounting for GPU time, power, cooling, and capital amortization — is somewhere between $20 and $40 per million tokens at current efficiency levels.

Every token generated at current prices loses money. OpenAI is subsidizing its customers.

Why This Is Different from Cloud Computing's Early Days

The comparison that OpenAI optimists reach for is early AWS. Amazon launched AWS in 2006 at prices that didn't cover costs, built scale, drove down costs through infrastructure investment and efficiency improvements, and eventually turned cloud computing into a margin machine. Why can't OpenAI execute the same playbook?

Three reasons.

First, the compute efficiency curve for AI inference is less favorable than the hardware efficiency curve for general compute. AWS's costs fell as Moore's Law continued delivering smaller, faster, cheaper chips. The efficiency gains in AI inference are real — Nvidia's Blackwell architecture delivers roughly 4x the inference throughput of Hopper per dollar — but the models are also getting larger and more capable faster than the hardware is getting cheaper. GPT-4 required dramatically more compute than GPT-3. GPT-5 will require dramatically more than GPT-4. The frontier keeps moving.

Second, the competitive dynamics are different. AWS competed against enterprises running their own data centers — a fragmented, slow-moving incumbent. OpenAI competes against Google, Microsoft, Anthropic, Meta, Mistral, and a growing list of open-source alternatives. Google has structural advantages in TPU infrastructure. Meta is releasing frontier-class models as open source. The price competition in AI APIs is already fierce and will intensify. OpenAI cannot simply wait for competitors to price rationally.

Third, the customer base has different price sensitivity. Enterprise AWS customers in 2008 were saving money compared to running their own infrastructure. OpenAI's enterprise customers are frequently paying for capabilities that didn't exist before and replacing processes that were handled by human labor. The value proposition is real, but the price-to-cost relationship is still being established. If OpenAI raises prices dramatically to approach cost-coverage, it accelerates the migration to cheaper alternatives.

What This Means for API-Dependent Businesses

If you're building a product on top of OpenAI's API — or Anthropic's, or Google's — this should keep you up at night for reasons that go beyond vendor risk.

The immediate risk is obvious: if OpenAI fails or dramatically curtails operations, you need an alternative. Dependency on a single API provider is a standard business continuity concern. The less obvious risk is what happens to pricing when the subsidy ends.

Current AI API prices are not market prices. They are subsidized prices, enabled by tens of billions in venture and strategic investment, designed to drive adoption. When — not if — the economics require prices to move toward cost coverage, the adjustment will be significant.

Consider the math for a mid-size AI-native startup currently spending $500,000 per month on inference costs. If true cost coverage requires 2x current pricing, that's a sudden $6 million annual increase in cost of goods sold. For a company with $5 million in revenue, that's existential. For a company with $20 million in revenue and 70% gross margins, it wipes out profitability entirely.

The accounting implications are immediate and material. Businesses that have built financial models, raised capital, and made hiring decisions based on current AI API pricing have baked a subsidy into their unit economics. That subsidy does not appear on their financial statements. Their investors may not realize it exists. When it ends, the impact will show up as a sudden deterioration in gross margins that is very difficult to explain without acknowledging that the model was built on an unsustainable foundation.

The Real Question: Who Pays?

Ultimately, somebody has to pay for AI inference. The options are limited.

Consumers pay more. OpenAI's ChatGPT Plus subscribers pay $20/month. If true cost-coverage requires $50/month, what happens to the subscriber base? Early evidence from streaming services suggests significant churn at price increases above 20-30%. The consumer AI market is not proven to be price-inelastic.

Enterprise customers pay more. This is more plausible. Enterprise software buyers are accustomed to significant price increases, and the switching costs for deeply integrated AI systems are real. But enterprise AI procurement is increasingly sophisticated, and buyers are already demanding contractual protections against pricing changes.

Compute costs fall fast enough. This is the hope. If Nvidia's next architecture delivers 10x the inference efficiency, if distillation and quantization continue improving, if inference optimization research unlocks significant efficiency gains, the math could improve dramatically without price increases. This is possible. It is not guaranteed. Betting your business on it is speculation.

Investors keep subsidizing. This is the current reality. OpenAI raised $40 billion. Saudi Arabia's PIF is a significant participant. The assumption that patient capital will continue funding losses until the economics work is not insane — but it's a geopolitical and financial bet that no business should make implicitly without acknowledging it explicitly.

The Accounting Question Nobody Is Asking

Here's the angle that should concern CFOs and their auditors: if you are building a business whose unit economics depend on subsidized input pricing, what are your disclosure obligations?

Under ASC 275 (Risks and Uncertainties), entities are required to disclose significant concentrations of risk when it is reasonably possible that those risks could result in a near-term severe impact on financial performance. A business that sources a primary input from a provider that is losing money at scale — and whose pricing doesn't cover costs — has a concentration of risk that is directly material to its own economic viability.

Most companies building on AI APIs are not making this disclosure. They are not flagging the pricing subsidy as a risk factor. Their auditors are not requiring it. This creates a gap between the financial picture presented to investors and the underlying economic reality.

The SEC hasn't issued guidance on this. FASB hasn't addressed it. But the conceptual framework is clear: material risks to the business must be disclosed. The dependency on subsidized AI infrastructure is a material risk for any company where AI API costs represent a significant portion of cost of goods sold.

What Businesses Should Actually Do

None of this means AI is a bad bet. The technology is real. The productivity gains are real. The long-term disruption is real. But the current pricing environment is not a permanent feature of the landscape, and businesses that treat it as such are making a category error.

Three things every AI-dependent business should be doing right now:

Build a multi-provider architecture. The ability to route inference to multiple providers — OpenAI, Anthropic, Google, and potentially open-source models running on self-managed infrastructure — is both a business continuity measure and a pricing negotiation tool. Single-provider dependency is a risk that can be engineered away.

Stress-test your unit economics at 2x and 3x current API pricing. If your business doesn't survive a doubling of inference costs, you don't have a business — you have a bet on subsidized pricing continuing indefinitely. That's not a business model. Know what your gross margins look like under realistic pricing scenarios and plan accordingly.

Understand the open-source alternative curve. Meta's Llama models are now competitive with GPT-4-level performance for many use cases. Running open-source models on owned or leased compute eliminates API pricing risk. The operational complexity is real — you're running infrastructure instead of calling an API — but for high-volume applications, the economics increasingly favor the self-hosted approach.

The businesses that will thrive in the post-subsidy AI era are the ones that built real moats — data advantages, workflow integration, switching costs — rather than the ones whose entire value proposition was built on cheap inference. The cheap inference era is ending. It was always going to end.

OpenAI losing $14 billion in 2026 isn't just a financing story. It's a signal that the economics of this industry are still unsettled in fundamental ways. The companies paying attention to that signal will be the ones standing when the dust settles.

Related Research

AI & Business Models8 min read

The SaaS Extinction Event: When AI Agents Replace Your $7,500/Year Software Stack

McKay Wrigley cancelled 7 B2B SaaS subscriptions in a single weekend after building replacements with Claude. He's not a...

Stay Current

New analysis on digital asset infrastructure, agent economics, and institutional crypto — delivered when published.

Grant Stellmacher, CPA
Finance Architect — Anchorage Digital · CPA Wisconsin #28430-1 · CPA Utah #14018703-2601