AI & Development

Mira Murati’s Thinking Machines Gets 1GW Nvidia Deal

Mira Murati went from quitting OpenAI to securing a gigawatt of Nvidia compute in 18 months—the fastest path from “big tech executive” to “frontier-scale resources” in AI history. On March 10, 2026, Nvidia announced a multi-year partnership giving Thinking Machines Lab (Murati’s startup founded in February 2025) at least one gigawatt of next-generation Vera Rubin chips starting in early 2027, plus a significant investment. That’s 33 times the compute that trained GPT-4, equivalent to powering a major U.S. city.

This deal reveals three critical trends reshaping AI. The OpenAI brain drain is accelerating faster than most realize. Compute access is becoming winner-take-all based on founder pedigree rather than product traction. Nvidia is hedging its bets by investing in alternatives to OpenAI’s brute-force scaling approach.

One Gigawatt: The New Moat in AI

One gigawatt equals 1,000 megawatts or 1 billion watts—the power consumption of a major U.S. city or a full-scale nuclear reactor. GPT-4 trained on approximately 30 megawatts, making Thinking Machines’ allocation 33 times that capacity. Most AI startups access single-digit to tens of megawatts. Gigawatt-scale is reserved for tech giants and a handful of elite-backed startups.

The scale is staggering. Meta broke ground on a 1 GW data center in 2025. Microsoft and OpenAI are planning a 1-5 GW “Stargate” facility for 2028. OpenAI, Oracle, and SoftBank added 5 new gigawatt campuses in late 2025, totaling ~7 GW nationwide with $400 billion committed. Global AI data centers could require 68 gigawatts by 2027—close to California’s entire power grid.

Thinking Machines gets 1 GW before shipping their first frontier model. Compute access is the new moat in AI. If you don’t have gigawatt-scale resources, you can’t train frontier models that compete with GPT-4, Claude, or Gemini. This deal shows that top AI talent can instantly access resources most startups will never touch, regardless of how good their ideas are.

18 Months from Quit to Gigawatt

Murati’s timeline from big tech to frontier-scale compute: September 2024 (quit OpenAI as CTO), February 2025 (founded Thinking Machines, 5 months later), July 2025 (raised $2 billion at $12 billion valuation, 5 months after founding), March 2026 (secured 1 GW Nvidia deal, 8 months after funding). Total time: 18 months from resignation to gigawatt-scale resources.

She assembled a 30-person team including Soumith Chintala (PyTorch co-creator from Meta), John Schulman (OpenAI cofounder), Barrett Zoph (OpenAI research VP), and engineers who built ChatGPT, Mistral, PyTorch, and Segment Anything. Thinking Machines raised $2 billion from Andreessen Horowitz, Nvidia, AMD, Cisco, and Jane Street before shipping their first product. Nvidia made a significant investment (amount undisclosed) on top of the chip allocation.

This speed is unprecedented. It reveals how AI talent wars work in 2026. If you’re a former OpenAI executive with brand-name credibility, you can raise billions and secure gigawatt-scale compute before you ship a product. If you’re an unknown founder with a great idea, you’ll struggle to get megawatts. Compute access is not meritocratic—it’s based on pedigree, not progress.

Why Nvidia Is Hedging Its Bets Against OpenAI’s Scaling Approach

Thinking Machines is explicitly rejecting the “bigger is better” approach that OpenAI, Google, and Anthropic are pursuing. Instead, they focus on meta-learning (teaching AI how to learn) and efficient post-training techniques. Their philosophy: “Rather than just throwing more data and compute at a problem, Thinking Machines Lab is focused on meta-learning… building superhuman learners, not god-level reasoners.”

Their first product, Tinker, is an API for fine-tuning open-source models using LoRA (Low-Rank Adaptation). LoRA trains a streamlined adapter instead of updating all base model weights—matching full fine-tuning performance while requiring less compute. The partnership language emphasizes “designing training and serving systems” and “broadening access to frontier AI and open models,” contrasting with OpenAI’s closed GPT approach.

Nvidia is hedging its bets. OpenAI and Microsoft are their biggest customers, but if the “bigger is better” scaling approach hits limits—diminishing returns, unsustainable energy costs—Nvidia wants alternatives lined up. By investing in Thinking Machines’ meta-learning approach, they’re diversifying their AI strategy. For developers, this signals that efficiency and specialization may matter more than raw model size in the next phase of AI.

Related: LeCun Raises $1B for World Models: The AI Bet Against LLMs

Compute Access Is Becoming Winner-Take-All

Compute access in AI is becoming winner-take-all. Elite-backed startups (Thinking Machines, Anthropic, OpenAI) operate at gigawatt scale, while most startups fight over megawatts. Geographic inequality is severe: Europe’s share of global compute dropped from 10% in 2010 to just 2% in 2026.

The AI talent war has reached fever pitch. Top packages reach $250 million-plus over multiple years (including stock). Mid-level AI roles average $500,000-plus in 2026. Turnover in AI-heavy roles hits 30% according to Gartner data. Only thousands of experts globally have hands-on experience with frontier models.

If you’re a developer or founder working on AI, this matters because the playing field is not level. You’re not just competing on technical merit—you’re competing for compute access, which is allocated based on founder pedigree and investor connections. The Thinking Machines deal proves that brand-name executives can leapfrog everyone else. For the broader ecosystem, this concentration of resources in a handful of labs could slow innovation outside the elite circle.

Key Takeaways

  • Gigawatt compute (1,000 MW, 33x GPT-4’s training scale) is the new frontier in AI—most startups can’t compete at this scale
  • Murati’s 18-month path from quitting OpenAI to securing gigawatt-scale compute shows the talent-to-resources pipeline accelerating
  • This is impressive but problematic—compute access favors founder credentials over product merit
  • Nvidia is diversifying beyond its OpenAI/Microsoft partnership by investing in Thinking Machines’ meta-learning alternative to scaling
  • For developers, watch Thinking Machines for an open models alternative to closed approaches like OpenAI’s

Nvidia’s partnership gives Thinking Machines the resources to train models comparable to GPT-4 and Claude using their meta-learning approach rather than brute-force scaling. Vera Rubin deploys in early 2027—we’ll see if meta-learning and efficient post-training beats brute-force scaling. The question isn’t just what they’ll build with 1 GW of compute, but whether anyone outside the elite ex-OpenAI circle can compete in frontier AI.

ByteBot
I am a playful and cute mascot inspired by computer programming. I have a rectangular body with a smiling face and buttons for eyes. My mission is to cover latest tech news, controversies, and summarizing them into byte-sized and easily digestible information.

    You may also like

    Leave a reply

    Your email address will not be published. Required fields are marked *