Home About Projects Blog Subscribe Login

The Hidden Economics of AI Training Runs

A single GPT-4 training run costs $100M+. The hyperscalers don't want you doing the math on who's subsidizing whom. Here's what the unit economics actually look like—and why they matter for every AI product.

OpenAI spent over $100 million training GPT-4. Anthropic likely spent similar amounts on Claude 3. Google, Meta, and xAI are all burning nine-figure budgets on their frontier models. The numbers are staggering—and they reveal an economic model that doesn't add up.

Here's the uncomfortable truth: the unit economics of foundation model training are broken. And understanding why matters for everyone building on top of these models—because the subsidy won't last forever.

The Training Cost Iceberg

When people talk about "$100M training runs," they're usually only counting compute. But that's just the tip of the iceberg. Let's break down the real cost structure:

Add it all up and you're looking at $200-300M per frontier model generation. And that's for a model that becomes obsolete in 12-18 months when the next generation ships.

The Revenue Side Doesn't Math

Now let's look at revenue. OpenAI is rumored to be doing $3-4B in ARR. Impressive—until you realize they're spending:

That's $1.3-1.6B in direct costs against $3-4B revenue. Margins are razor-thin—and that's before accounting for sales, marketing, compliance, and everything else that goes into running a business.

The path to profitability? Raise prices or cut costs. But they can't really do either:

Who's Really Paying for This?

So if the economics don't work, who's subsidizing the AI boom? Three groups:

1. Venture Capital

OpenAI raised $13B+ (mostly from Microsoft). Anthropic raised $7B+. Inflection raised $1.5B before pivoting. VCs are essentially paying for the training runs in exchange for equity in a future where these companies dominate.

This works as long as investors believe in a path to massive exits. But if margins stay thin and competition stays fierce, that story gets harder to tell.

2. Hyperscalers (Cloud Providers)

Microsoft gave OpenAI Azure credits as part of their investment. Google is subsidizing DeepMind and Gemini. Amazon backs Anthropic with AWS credits.

Why? Because AI is driving cloud consumption. Every startup using GPT-4 is paying Microsoft for compute. Every company fine-tuning Claude is paying Amazon. The model providers might break even, but the cloud providers are printing money on inference.

3. End Users (Indirectly)

You're paying $20/month for ChatGPT Plus. Microsoft is charging you $30/month for Copilot. But these prices don't cover the true cost—they're loss leaders designed to lock in users and gather training data.

The real monetization comes later: enterprise contracts, API usage at scale, and proprietary models built on the data you generate. You're not the customer—you're the training set.

What This Means for Builders

If you're building on top of foundation models, these economics have huge implications:

Expect Price Volatility

The current pricing is artificially low. As VCs demand profitability and cloud subsidies shrink, prices will rise. Don't assume today's API costs will hold for three years.

Diversify Model Dependencies

OpenAI might dominate today, but economic pressure could force consolidation. Build your product to work across multiple providers. Use routers, fallbacks, and abstraction layers so you're not locked into one vendor's pricing power.

Optimize for Inference Cost

Training costs are someone else's problem. Your problem is inference. Every API call compounds. Use smaller models where possible, cache aggressively, and batch requests. The companies that win on AI will be the ones that optimize cost per task, not raw capability.

Watch the Open Source Alternative

Llama 3, Mistral, and other open models are closing the gap. If you can fine-tune a 70B open model to match GPT-4 on your specific task, you've just eliminated your inference dependency entirely. The economics of self-hosting start to make sense at scale.

The End Game

One of three things happens:

1. Prices Rise
Model providers find a path to profitability by charging what training and inference actually cost. Enterprise customers pay it (they have no choice). Consumer products get more expensive or disappear.

2. Consolidation
Most model providers can't sustain the burn rate. They get acquired by hyperscalers or shut down. OpenAI becomes a Microsoft product. Anthropic becomes an Amazon service. Competition dies, prices rise anyway.

3. Commoditization
Open models catch up fully. Training costs drop (more efficient architectures, cheaper compute). Inference gets faster and cheaper (quantization, distillation, dedicated hardware). Foundation models become a low-margin infrastructure business—profitable, but boring.

I'm betting on a mix of 2 and 3. The frontier will consolidate (only a few players can afford $300M training runs). But the long tail will commoditize (good-enough models get cheap and open).

Either way, the subsidy era is temporary. The companies building on AI today need to plan for a world where API calls cost 3-5x what they do now. If your unit economics don't work at that price, you're building on borrowed time.

The Bottom Line

AI training runs are expensive because they have to be—this is the cost of pushing the frontier. But the current pricing model is unsustainable. VCs and cloud providers are footing the bill for now, but they won't forever.

If you're building on AI: optimize for cost, diversify dependencies, and prepare for price increases. The free lunch is ending.

And if you're investing in AI: follow the money. The real winners won't be the model providers—they're stuck in a margin-crushing race. The winners will be the infrastructure players (NVIDIA, hyperscalers) and the application layer companies that figure out how to deliver value even when API costs rise.

The AI gold rush is real. But like every gold rush, the people selling picks and shovels makes more money than the miners.


Follow the journey

Subscribe to Lynk for daily insights on AI strategy, cybersecurity, and building in the age of AI.

Subscribe →