NVIDIA's $40k H100s Finally Pushed Someone Over the Edge

OpenAI Logo

OpenAI is dropping around $10 billion on custom chips with Broadcom because NVIDIA's pricing has gotten completely insane. H100s cost around $40k each with months-long wait times. If you can even get them.

When Broadcom's CEO casually dropped that they just landed a "$10+ billion order from a new customer," everyone knew it was OpenAI. Who else is desperate enough to spend that much on custom silicon? The answer: anyone burning hundreds of thousands per day just to keep ChatGPT running. OpenAI's total AI training and inference costs could hit $7 billion in 2025.

Broadcom's stock jumped 16% because investors finally realized someone found a way to compete with NVIDIA's monopoly. About fucking time. NVIDIA has been charging whatever they want because they had zero competition.

The plan is inference-focused chips starting in 2026. Unlike training chips that need massive parallel horsepower, inference chips optimize for cost and efficiency when you're serving millions of users. Think "make ChatGPT responses cheap" not "train GPT-6."

Here's the math that drove this decision: OpenAI is projecting massive cash burn through 2029 - probably over $100 billion more than their previous estimates. When you're burning money that fast on compute costs, spending $10 billion upfront to cut your operational expenses starts making sense.

This isn't just OpenAI being dramatic. Google built TPUs, Amazon has Trainium, and Apple designed their M-series specifically to avoid paying Intel's markup. AWS Trainium and Google TPU claim significant cost savings per billion parameters versus NVIDIA solutions. Every major tech company eventually hits the point where building custom silicon beats paying someone else's margin.

NVIDIA owns most of the AI accelerator market, largely due to CUDA software dominance. But custom AI chip development by major tech companies is increasingly challenging this monopoly position.

The custom AI accelerator market could reach $45 billion by 2027. TSMC's advanced process nodes are becoming critical battlegrounds, with foundry capacity constraints forcing companies to book production years in advance.

H100s are the backbone of basically every AI company's infrastructure. NVIDIA's gross margins on H100s probably exceed 70%, creating massive incentives for customers to develop alternatives. But when your monopoly pricing pushes customers to spend $10 billion on alternatives, maybe you got too greedy.

What Could Go Wrong with This $10B Bet?

The Broadcom-OpenAI deal sounds great on paper, but custom chips are where companies go to burn money and miss deadlines. The promised 2-5x performance improvements assume everything works perfectly. In reality, first-gen custom silicon usually sucks.

The Hard Reality of Custom Silicon

OpenAI's betting they can optimize for transformer inference patterns better than NVIDIA optimizes for everything. Maybe. But custom ASICs are a bitch to get right. You're locked into whatever architecture decisions you make in 2025, while NVIDIA keeps iterating every 12 months.

Sure, the chips could have:

  • Optimized memory for attention patterns (if you guess the memory access patterns correctly)
  • Custom token processing (until OpenAI changes their model architecture)
  • Integrated networking (that works with exactly one data center setup)
  • Power optimizations (that TSMC actually delivers on time)

Everyone Else Is Failing at This

Google's TPUs work because Google controls their entire stack. Meta's MTIA chips are still trying to catch up to H100s. Microsoft's Azure Maia processors launched a year late. Custom chips are where ambitious roadmaps go to die.

Amazon's Inferentia took three generations to become competitive with NVIDIA. That's 6+ years of development. OpenAI thinks they'll nail it in one shot by 2026.

The 2026 Timeline Is Delusional

Custom chip development takes 3-5 years minimum. The fact that Broadcom confidently announced 2026 means either:

  1. They've been secretly working on this for years
  2. They're using existing chip designs and calling them "custom"
  3. Someone's lying about the timeline

Broadcom knows networking and storage, not AI accelerators. Their VMware acquisition was about software, not silicon. This is like hiring a plumber to design your car engine.

The real kicker? They need TSMC's 3nm or 2nm process for competitive performance. Good luck getting foundry capacity. Apple, NVIDIA, and every other major tech company are already fighting for those slots. TSMC doesn't give a shit about your $10 billion if you can't guarantee multi-year volume commitments.

What Developers Actually Want to Know

Q

Is NVIDIA completely fucked now?

A

Not yet, but they should be sweating. H100s still cost $25k-40k each with months-long wait times. If OpenAI's chips actually work, every other AI company will want their own custom silicon. NVIDIA's monopoly pricing pushed their biggest customers to build alternatives.

Q

Will this make AI APIs cheaper for startups?

A

Maybe in 3-4 years, if at all.

OpenAI isn't building these chips to cut your API costs

GPT/comments/15qu10a/chatgpt_costs_openai_700000_per_day/) on compute. Any savings will probably go toward training bigger models, not cheaper pricing.

Q

Should I sell my NVIDIA stock?

A

Fuck no. Custom chips take years to actually work, and most companies don't have $10 billion lying around. NVIDIA will keep printing money from everyone else while OpenAI figures out if their chips actually work. Plus, training still needs H100s.

Q

What happens when these chips inevitably suck?

A

OpenAI will quietly keep buying H100s while claiming their custom chips are "ramping up." Meta's MTIA chips took years to match NVIDIA performance. First-gen custom silicon always disappoints.

Q

Why Broadcom? They don't do AI chips.

A

Because Intel and AMD would take forever and leak everything to competitors. Broadcom does custom ASICs and keeps their mouth shut. Sometimes you pick the plumber who shows up, not the one with the best Yelp reviews.

Q

When will this actually matter?

A

2026 if you're optimistic, 2028 if you're realistic. Custom chips are where good intentions go to die. Even if it works perfectly, you're looking at years before it affects anything outside OpenAI's data centers.

Related Tools & Recommendations

news
Similar content

OpenAI & Broadcom's $10B AI Chip Deal: Ditching NVIDIA

$10B deal with Broadcom because they're sick of paying NVIDIA's ridiculous prices

OpenAI/ChatGPT
/news/2025-09-05/broadcom-openai-10b-chip-partnership
100%
news
Similar content

Broadcom Lands $10B OpenAI AI Chip Deal: Custom Silicon by 2026

OpenAI finally decided Nvidia's pricing is bullshit, custom silicon shipping 2026

OpenAI/ChatGPT
/news/2025-09-05/broadcom-openai-10b-chip-deal
90%
news
Recommended

Claude AI Can Now Control Your Browser and It's Both Amazing and Terrifying

Anthropic just launched a Chrome extension that lets Claude click buttons, fill forms, and shop for you - August 27, 2025

chrome
/news/2025-08-27/anthropic-claude-chrome-browser-extension
60%
news
Similar content

AMD UDNA Flagship GPU: Challenging NVIDIA with New Architecture

UDNA Architecture Promises High-End GPUs by 2027 - If They Don't Chicken Out Again

OpenAI ChatGPT/GPT Models
/news/2025-09-01/amd-udna-flagship-gpu
59%
news
Similar content

Alibaba's AI Chip: China's Answer to Nvidia H20s Ban

China's Backup Plan for When Uncle Sam Cuts Off the Chips

NVIDIA GPUs
/news/2025-08-29/alibaba-ai-chip-nvidia-void
57%
news
Similar content

Alibaba Unveils AI Chip: Challenging Nvidia's China Dominance

Chinese tech giant launches advanced AI inference processor as US-China chip war escalates

OpenAI ChatGPT/GPT Models
/news/2025-08-31/alibaba-ai-chip-nvidia-challenge
54%
news
Similar content

Mistral AI: Europe's €12B AI Champion Secures €2B Funding

French Startup Hits €12B Valuation While Everyone Pretends This Makes OpenAI Nervous

/news/2025-09-03/mistral-ai-2b-funding
50%
news
Similar content

Alibaba's AI Chips: China's Answer to NVIDIA Sanctions

Alibaba Built Their Own AI Chips After Getting Cut Off From the Good Stuff

Samsung Galaxy Devices
/news/2025-08-30/alibaba-ai-chip-nvidia-void
50%
tool
Recommended

Ollama Production Deployment - When Everything Goes Wrong

Your Local Hero Becomes a Production Nightmare

Ollama
/tool/ollama/production-troubleshooting
49%
compare
Recommended

Ollama vs LM Studio vs Jan: The Real Deal After 6 Months Running Local AI

Stop burning $500/month on OpenAI when your RTX 4090 is sitting there doing nothing

Ollama
/compare/ollama/lm-studio/jan/local-ai-showdown
49%
news
Similar content

Nvidia's $45B Earnings Test: AI Chip Tensions & Tech Market Impact

Wall Street set the bar so high that missing by $500M will crater the entire Nasdaq

GitHub Copilot
/news/2025-08-22/nvidia-earnings-ai-chip-tensions
49%
news
Similar content

Alibaba's RISC-V AI Chip: Breakthrough or Hype?

Alibaba announces "breakthrough" RISC-V chip that still can't train models, promises Samsung's entire yearly revenue in investments

OpenAI ChatGPT/GPT Models
/news/2025-09-01/alibaba-ai-chip-breakthrough
47%
news
Similar content

Gemini 2.0 Flash vs. Sora: Latest AI Model News & Updates

Gemini 2.0 vs Sora: The race to burn the most venture capital while impressing the fewest users

General Technology News
/news/2025-08-24/ai-revolution-accelerates
45%
news
Similar content

Nvidia Spectrum-XGS: Revolutionizing GPU Networking for AI

Enterprise AI Integration Brings Advanced Reasoning to Business Workflows

GitHub Copilot
/news/2025-08-22/nvidia-spectrum-xgs-networking
43%
news
Similar content

NVIDIA Earnings: AI Market's Crucial Test Amid Tech Decline

Wall Street focuses on NVIDIA's upcoming earnings as tech stocks waver and AI trade faces critical evaluation with analysts expecting 48% EPS growth

GitHub Copilot
/news/2025-08-23/nvidia-earnings-ai-market-test
43%
integration
Recommended

PyTorch ↔ TensorFlow Model Conversion: The Real Story

How to actually move models between frameworks without losing your sanity

PyTorch
/integration/pytorch-tensorflow/model-interoperability-guide
43%
news
Similar content

OpenAI & Anthropic Reveal Critical AI Safety Testing Flaws

Two AI Companies Admit Their Safety Systems Suck

OpenAI ChatGPT/GPT Models
/news/2025-08-31/ai-safety-testing-concerns
42%
news
Similar content

NVIDIA AI Chip Sales Cool: Q2 Misses Estimates & Market Questions

Q2 Results Miss Estimates Despite $46.7B Revenue as Market Questions AI Spending Sustainability

/news/2025-08-28/nvidia-ai-chip-slowdown
42%
news
Similar content

NVIDIA & Trump Negotiate China AI Chip Sales: Blackwell GPUs

Scaled-Down Blackwell GPUs Could Resume Limited China Trade

NVIDIA AI Chips
/news/2025-08-29/nvidia-china-blackwell-negotiations
42%
news
Similar content

Jensen Huang: NVIDIA's Quantum Computing Future & AI Hybrid Systems

NVIDIA CEO makes bold claims about quantum-AI hybrid systems, because of course he does

Samsung Galaxy Devices
/news/2025-08-30/nvidia-quantum-computing-bombshells
42%

Recommendations combine user behavior, content similarity, research intelligence, and SEO optimization