NVIDIA and OpenAI Forge $100B Alliance to Power the Next AI Revolution

The collaboration marks a landmark shift in AI hardware development, with NVIDIA stepping in as both supplier and strategic investor. By aligning its product roadmap with OpenAI’s next-generation models, the partnership tackles critical hurdles — from supply chain and power constraints to regional deployment logistics — setting the stage to scale AI innovation on a truly global level.
Sept. 24, 2025
9 min read

Key Highlights

  • OpenAI and NVIDIA's partnership aims to deploy over 10 GW of compute capacity, starting with 1 GW in late 2026, to support next-generation AI models.
  • NVIDIA plans to invest up to $100 billion, with an initial $10 billion commitment, to finance hardware deployment without controlling voting rights, aligning incentives for long-term growth.
  • The new Vera Rubin platform and NVL144 systems will enable large context windows and multimodal workloads, emphasizing hardware-software co-design for AI training and inference.
  • This initiative diversifies OpenAI's infrastructure sources beyond Microsoft, Oracle, and others, reducing dependency on a single cloud provider and broadening hardware ecosystem integration.
  • Operational challenges include massive power requirements and site logistics, with deployment likely distributed across multiple campuses due to scale and infrastructure constraints.

The new strategic partnership between OpenAI and NVIDIA, formalized via a letter of intent in September 2025, is designed to both power and finance the next generation of OpenAI’s compute infrastructure, with initial deployments expected in the second half of 2026. According to the joint press release, both parties position this as “the biggest AI infrastructure deployment in history,” explicitly aimed at training and running OpenAI’s next-generation models. 

At a high level:

  • The target scale is 10 gigawatts (GW) or more of deployed compute capacity, realized via NVIDIA systems (comprising millions of GPUs). 
  • The first phase (1 GW) is slated for the second half of 2026, built on the forthcoming Vera Rubin platform. 
  • NVIDIA will progressively invest up to $100 billion into OpenAI, contingent on deployment of capacity in stages. 
  • An initial $10 billion investment from NVIDIA is tied to the execution of a definitive purchase agreement for the first gigawatt of systems. 
  • The equity stake NVIDIA will acquire is described as non-voting / non-controlling, meaning it gives financial skin in the game without governance control. 

From a strategic standpoint, tying investment to capacity deployment helps OpenAI lock in capital and hardware over a long horizon, mitigating supply-chain and financing risk. With compute frequently cited as a binding constraint on advancing models, this kind of staged, anchored commitment gives OpenAI a more predictable growth path (at least in theory; that said, the precise economic terms and risk-sharing remain to be fully disclosed.)

Press statements emphasize that millions of GPUs will ultimately be involved, and that co-optimization of NVIDIA’s hardware with OpenAI’s software/stack will be a key feature of the collaboration. 

Importantly, this deal also fits into OpenAI’s broader strategy of diversifying infrastructure partnerships beyond any single cloud provider. Microsoft remains a central backer and collaborator, but this NVIDIA tie-up further broadens OpenAI’s compute base, complementing other announced partnerships (e.g. with Oracle, SoftBank, and Stargate). 

This deal also marks a strategic shift for NVIDIA: rather than being purely a vendor of chips, it becomes a strategic investor in an anchor customer whose growth directly drives GPU demand. This alignment tightens the coupling between NVIDIA’s roadmap, its software ecosystem, and real-world end deployments.

In a CNBC interview, NVIDIA CEO Jensen Huang characterized the initiative thus:

“This is the biggest AI infrastructure project in history. This partnership is about building an AI infrastructure that enables AI to go from the labs into the world.” 

NVIDIA’s press materials also add that this new AI infrastructure will deliver “a billion times more computational power” than the first DGX system Huang delivered to OpenAI in 2016 — a rhetorical contrast intended to highlight the scale leap. 

Significant Architectural Upgrade for OpenAI

NVIDIA’s Vera Rubin family and the new NVL144 CPX systems represent a deliberate architectural pivot toward ultra-dense, rack-scale platforms optimized for very long context windows, multimodal workloads, and generative video. NVIDIA’s Rubin CPX announcement frames the platform around “million-token” inference/prefill use cases and emphasizes very large on-rack memory and extreme cross-rack bandwidth — the Vera Rubin NVL144 CPX rack is advertised as delivering roughly 8 exaflops of NVFP4 AI compute, ~100 TB of high-speed memory per rack and on the order of 1.7 PB/s of aggregate memory bandwidth, with an overall performance uplift NVIDIA says is about 7.5× versus the prior GB300 NVL72 systems. 

Technically, the Rubin family also introduces a disaggregated approach to long-context inference: NVIDIA is shipping both a compute-optimized Rubin CPX (reported at ~30 PFLOPs NVFP4 with 128 GB of GDDR7 per socket in some configurations) and higher-bandwidth Rubin GPUs (larger HBM configurations) so different phases of inference (prefill/context vs. generate/decode) can be mapped to the most appropriate silicon. This split — together with next-generation NVLink/NVSwitch fabrics, a new NVLink-144 switch and enhanced silicon-photonics and NICs referenced in platform materials — is intended to deliver much higher effective context lengths and throughput when combined with software optimizations. 

For OpenAI, the practical implication is hardware + software co-design at scale: Rubin-era GPUs, NVL144 NVLink/NVSwitch fabrics and accelerated networking will be tuned to OpenAI’s training and inference pipelines, and CUDA/SDK roadmap alignment should ease integration and performance tuning. Public coverage and NVIDIA materials explicitly call out co-optimization and rack-scale system designs intended for million-token workloads. 

There are operational and deployment consequences worth flagging. The NVL144 CPX racks are being positioned for production shipments in late 2026 (NVIDIA’s public timetable), which aligns with the 1-GW first-stage timing announced in the NVIDIA–OpenAI LOI. Scaling the kind of capacity OpenAI and NVIDIA describe will likely require distributed deployment across multiple campuses and providers — multiple independent reporting outlets and systems-level analyses emphasize the platform’s rack-scale nature and the practical limits (power, cooling, site procurement) that make a single-campus 10-GW buildout unlikely and operationally risky. Treat the “distributed campuses” statement as an informed inference from platform design and the industry’s power/site realities rather than a line from an NDA or definitive filing. 

Sam Altman, cofounder and CEO of OpenAI says:

“Everything starts with compute. Compute infrastructure will be the basis for the economy of the future, and we will utilize what we’re building with NVIDIA to both create new AI breakthroughs and empower people and businesses with them at scale.”

It’s clear that OpenAI is betting on NVIDIA to serve as the engine driving the next wave of AI hardware improvements. In turn, NVIDIA is betting on OpenAI to strengthen its position as the most visible AI application provider, deepening customer lock-in. The partnership also helps NVIDIA defend its market share against hyperscalers or startups developing custom silicon, by demonstrating that the fastest proven route to frontier-scale models runs on NVIDIA systems — with OpenAI’s deployments as living proof.

Greg Brockman, Co-founder and President of OpenAI, said:

“We’ve been working closely with NVIDIA since the early days of OpenAI. We’ve utilized their platform to create AI systems that hundreds of millions of people use every day. We’re excited to deploy 10 gigawatts of compute with NVIDIA to push back the frontier of intelligence and scale the benefits of this technology to everyone.”

Potential Problems

Reporting from Reuters and other outlets has underscored the possibility that this arrangement could attract antitrust and competition scrutiny. Because NVIDIA would be both the dominant supplier of AI accelerators and a direct investor in OpenAI, regulators in the U.S. and abroad may see conflicts of interest or barriers to fair competition. This scrutiny could lead to delays, conditions, or even significant modifications to the deal. In a worst-case scenario, agencies could block the agreement outright if they judge it to concentrate too much market power in a single vendor–customer axis.

The power challenge is equally formidable. Deploying 10 GW of AI infrastructure does not only mean racks of GPU servers; it implies building or securing entire new power plants, substations, and transmission infrastructure. Even with the recent federal executive order streamlining permitting for AI-related infrastructure and data centers framed as national security priorities, local and regional siting processes remain unpredictable. Community opposition, environmental review, or transmission-line bottlenecks can add years to project timelines. Each gigawatt represents a major utility-scale deployment, and stacking ten of them is unprecedented in data center history. Delays in grid interconnection or local permitting are therefore a material risk to the announced timeline.

The Bigger Picture

At its heart, the OpenAI–NVIDIA agreement is more than a procurement plan: it’s a bold wager on the data center as the engine room of the AI economy. If the parties can navigate the hurdles of regulation, power sourcing, and deployment at unprecedented scale, this partnership could redefine what global infrastructure looks like in the AI era.

The risks are real, but so is the ambition. A 10-gigawatt roadmap signals that frontier AI is no longer just about research breakthroughs: it’s about building civilizational-scale infrastructure. For the industry, this story may mark a turning point: where the alignment of hardware, capital, and vision has the potential to reshape both the market for compute and the trajectory of human–machine collaboration.

 

At Data Center Frontier, we talk the industry talk and walk the industry walk. In that spirit, DCF Staff members may occasionally use AI tools to assist with content. Elements of this article were created with help from OpenAI's GPT5.

 
Keep pace with the fast-moving world of data centers and cloud computing by connecting with Data Center Frontier on LinkedIn, following us on X/Twitter and Facebook, as well as on BlueSky, and signing up for our weekly newsletters using the form below.

About the Author

David Chernicoff

David Chernicoff

David Chernicoff is an experienced technologist and editorial content creator with the ability to see the connections between technology and business while figuring out how to get the most from both and to explain the needs of business to IT and IT to business.

Matt Vincent

A B2B technology journalist and editor with more than two decades of experience, Matt Vincent is Editor in Chief of Data Center Frontier.

Sign up for the Data Center Frontier Newsletter
Get the latest news and updates.
VideoFlow/Shutterstock.com
Source: VideoFlow/Shutterstock.com
Sponsored
Justine Belyus, Executive Director of Business Marketing for the Pennsylvania Department of Community & Economic Development, explains why Pennsylvania is ready to power the next...
panumas nikhomkhai/Shutterstock.com
Source: panumas nikhomkhai/Shutterstock.com
Sponsored
Benjamin Taylor of HT Materials Science outlines recent innovations in HVAC technology that offer cost-effective, efficient solutions to ease the strain data centers place on ...