OpenAI is dropping around $10 billion on custom chips with Broadcom because NVIDIA's pricing has gotten completely insane. H100s cost around $40k each with months-long wait times. If you can even get them.
When Broadcom's CEO casually dropped that they just landed a "$10+ billion order from a new customer," everyone knew it was OpenAI. Who else is desperate enough to spend that much on custom silicon? The answer: anyone burning hundreds of thousands per day just to keep ChatGPT running. OpenAI's total AI training and inference costs could hit $7 billion in 2025.
Broadcom's stock jumped 16% because investors finally realized someone found a way to compete with NVIDIA's monopoly. About fucking time. NVIDIA has been charging whatever they want because they had zero competition.
The plan is inference-focused chips starting in 2026. Unlike training chips that need massive parallel horsepower, inference chips optimize for cost and efficiency when you're serving millions of users. Think "make ChatGPT responses cheap" not "train GPT-6."
Here's the math that drove this decision: OpenAI is projecting massive cash burn through 2029 - probably over $100 billion more than their previous estimates. When you're burning money that fast on compute costs, spending $10 billion upfront to cut your operational expenses starts making sense.
This isn't just OpenAI being dramatic. Google built TPUs, Amazon has Trainium, and Apple designed their M-series specifically to avoid paying Intel's markup. AWS Trainium and Google TPU claim significant cost savings per billion parameters versus NVIDIA solutions. Every major tech company eventually hits the point where building custom silicon beats paying someone else's margin.
NVIDIA owns most of the AI accelerator market, largely due to CUDA software dominance. But custom AI chip development by major tech companies is increasingly challenging this monopoly position.
The custom AI accelerator market could reach $45 billion by 2027. TSMC's advanced process nodes are becoming critical battlegrounds, with foundry capacity constraints forcing companies to book production years in advance.
H100s are the backbone of basically every AI company's infrastructure. NVIDIA's gross margins on H100s probably exceed 70%, creating massive incentives for customers to develop alternatives. But when your monopoly pricing pushes customers to spend $10 billion on alternatives, maybe you got too greedy.