Designing the AI Factory: Cadence’s Sherman Ikemoto on Digital Twins, Power Reality, and the End of Guesswork

In this episode of the DCF Show, Cadence’s Sherman Ikemoto explains how digital twins, validated system models, and SLA-driven simulation are giving operators a clearer way to design and run GPU-heavy data centers, replacing guesswork with engineering discipline in the age of AI.
Jan. 20, 2026
10 min read

Key Highlights

  • Modern AI data centers are systems that require dense, interdependent design approaches beyond traditional rules of thumb.
  • Cadence's digital infrastructure elements, like the behaviorally accurate DGX SuperPOD model, enable pre-construction simulation to predict performance and stress scenarios.
  • Validated models with star ratings ensure confidence in outcomes, reducing safety margins and optimizing costs for high-density, high-performance environments.
  • Simulation and digital twins are crucial for understanding complex interactions between power, cooling, and IT systems, especially as densities increase and traditional tools fall short.
  • The integration of NVIDIA Omniverse into digital twin platforms supports extreme co-design, enabling real-time data exchange and early conflict detection across system domains.

The AI data center is no longer just a building full of racks. It is a system: dense, interdependent, and increasingly unforgiving of bad assumptions.

That reality sits at the center of the latest episode of The Data Center Frontier Show, where DCF Editor-in-Chief Matt Vincent sits down with Sherman Ikemoto, Senior Director of Product Management at Cadence, to talk about what it now takes to design an “AI factory” that actually works.

The conversation ranges from digital twins and GPU-dense power modeling to billion-cycle power analysis and the long-running Cadence–NVIDIA collaboration. But the through-line is simple: the industry is outgrowing rules of thumb.

As Ikemoto puts it, data center design has always been a distributed process. Servers are designed by one set of suppliers, cooling by another, power by another. Only at the end does the operator attempt to integrate those parts into a working system.

That final integration phase, he argues, has long been underserved by design tools. The risk shows up later, as downtime, cost overruns, or performance shortfalls.

Cadence’s answer is a new class of digital infrastructure: what it calls “DC elements,” validated building blocks that let operators assemble and simulate an AI factory before they ever pour concrete.

The DGX SuperPOD as a Digital Building Block

One of the most significant recent additions is a full behavioral model of NVIDIA’s DGX SuperPOD built around GB200 systems. This is not just a geometry file or a thermal sketch. It is a behaviorally accurate digital representation of how that system consumes power, moves heat, and interacts with airflow and liquid cooling.

In practice, that means an operator can drop a DGX SuperPOD element into a digital design and immediately see how it stresses the rest of the facility: power distribution, cooling loops, airflow patterns, and failure scenarios.

Ikemoto describes these as “DC elements”: digital building blocks that bridge the gap between suppliers and operators. Each element carries real performance data from the vendor, validated through a star-rating system that measures how closely the model matches real-world behavior.

The highest rating, five stars, requires strict validation and supplier sign-off. The GB200 DGX SuperPOD element reached that level, developed in close collaboration with NVIDIA.

The goal is not elegance in simulation. It is confidence in outcomes.

Designing to an SLA

Every data center is built to meet a set of service-level agreements, whether explicit or implied. Historically, uncertainty in design data forced engineers to pad those designs with large safety margins.

That padding shows up as overbuilt power systems, oversized cooling, and higher costs.

Ikemoto argues that GPU-dense environments are making that approach untenable. When racks move from 10–20 kW to 50–100 kW (and toward a future that could reach megawatt-class racks) guesswork becomes expensive and dangerous.

Simulation, paired with validated DC elements, allows designers to model directly to an SLA: uptime targets, thermal limits, power availability, and cost-performance thresholds. The more precise the input data, the smaller the required design margin.

In other words, better models let operators spend less money to get more reliability.

What “Behaviorally Accurate” Really Means

Calling a model “accurate” is easy. Proving it is harder.

Cadence validates every DC element against supplier data, using a star system that reflects depth of detail and level of verification. Five-star elements undergo the strictest validation and require supplier approval.

For extreme-density systems like GB200, that means validating not just steady-state power and thermal behavior, but how those systems respond dynamically—under load changes, partial failures, and localized stress.

That level of fidelity is what allows simulation to move from illustrative to operational.

The Real Bottleneck: Knowledge

When asked about the biggest bottleneck in AI data center buildouts, Ikemoto does not point first to utilities, transformers, or chillers. He points to knowledge.

The industry is moving from a world of gradual IT evolution to one of discontinuous jumps. Power density has multiplied in just a few years. Cooling has shifted from mostly air to increasingly liquid. Interactions between systems have become tighter and more fragile.

Traditional design tools and rules of thumb evolved for a slower world. They struggle to keep up with this rate of change.

Physical prototyping is too slow and too expensive to fill the gap. Virtual prototyping, i.e. simulation, offers a way to build that knowledge faster, safer, and at lower cost. Ikemoto compares the moment to what aerospace and automotive industries went through decades ago, when simulation became the backbone of design.

From CFD to a Digital Twin

Computational Fluid Dynamics has long been used in data centers to study airflow and heat transfer. But traditional CFD tools are slow, fragile, and built for general-purpose engineering, not for the specific realities of data centers.

Cadence’s Reality Digital Twin platform includes a custom CFD engine designed specifically for data center environments. It is optimized for airflow, liquid cooling, and hybrid systems, and built to run fast and reliably.

More importantly, it is embedded in a broader platform that supports design, commissioning, and operations. It can simulate next-generation AI factory architectures that mix air and liquid cooling, and critically, show how those systems interact.

That interaction is where many modern failures hide.

Extreme Co-Design and Omniverse

AI factories are too complex for power, cooling, IT layout, and operations to be designed in isolation. Changes in one domain ripple into the others.

NVIDIA calls this “extreme co-design”: designing all major systems together, continuously, to avoid late-stage conflicts.

Cadence is embedding NVIDIA Omniverse capabilities into its digital twin platform to support that approach. The vision is a shared simulation environment where different design tools exchange data in real time, conflicts are caught early, and optimization happens across domains, not in silos.

Looking further out, Ikemoto describes NVIDIA’s long-term vision of the AI factory as a largely autonomous system: robots installing, maintaining, and upgrading infrastructure, guided by a digital twin that acts as the factory’s operating system.

In that future, the digital twin is not just a design tool. It is the control layer for a living system.

The Twin After Day One

Too many digital twins stop at commissioning. Cadence’s Reality platform is designed to carry forward into operations.

Ikemoto says more than 2.5 million square feet of data center space (or nearly 700 data halls globally) already use digital twins operationally.

A model that begins in design becomes, after commissioning, a virtual replica of a specific physical facility. Connected to live monitoring systems, it can simulate future changes before they happen: IT upgrades, maintenance work, power reconfiguration.

Operators can test decisions in the twin before risking them in steel and silicon. The same visual model can be used to train people and eventually robots who will operate those facilities.

The shift is from digital twin as project artifact to digital twin as operational platform.

What Real AI Workloads Reveal

One of the most revealing parts of the conversation centers on what happens when real AI and machine-learning traces are run through simulation.

Some workloads create short, sharp bursts of power demand in localized areas. To be safe, many facilities provision power 20–30% above what they usually need, just to cover those spikes.

Most of that extra capacity sits unused most of the time. In an era where power is the primary bottleneck for AI factories, that waste matters.

By linking application behavior to hardware power profiles and then to facility-level power distribution, simulation can help smooth those spikes and reduce overprovisioning. It ties software design, IT design, and facility design into a single performance conversation.

That integration is where cost performance now lives.

Billion-Cycle Power Analysis

The conversation also touches on Cadence’s new Palladium Dynamic Power Analysis App, which can analyze billion-gate designs across billions of cycles with up to 97% accuracy.

What made that possible, Ikemoto says, is a combination of advances in emulation hardware and new algorithms that allow massive workloads to be profiled at realistic scale.

For data centers, the significance is indirect but profound. Better chip-level power modeling feeds better system-level models, which feed better facility-level designs.

The data center is, in Ikemoto’s words, “the ultimate system”: the place where all these components finally meet.

A Collaboration That Keeps Expanding

Cadence and NVIDIA have worked together for decades, with NVIDIA long relying on Cadence tools to design its chips.

What has changed is the scope of that collaboration. As accelerated computing reshapes infrastructure, their shared work has expanded from chips to systems to entire data centers.

The partnership now stretches from silicon to server to rack to AI factory. The common thread is simulation as a way to manage complexity before it becomes expensive.

Ikemoto frames it as a response to the digital transformation of the world. As more of life depends on compute, the infrastructure behind that compute has to be designed with the same rigor as the chips inside it.

The AI factory is not a metaphor anymore. It is an industrial system and like any industrial system, it needs blueprints that reflect reality.

 

At Data Center Frontier, we talk the industry talk and walk the industry walk. In that spirit, DCF Staff members may occasionally use AI tools to assist with content. Elements of this article were created with help from OpenAI's GPT5.

 
Keep pace with the fast-moving world of data centers and cloud computing by connecting with Data Center Frontier on LinkedIn, following us on X/Twitter and Facebook, as well as on BlueSky, and signing up for our weekly newsletters using the form below.

About the Author

Matt Vincent

A B2B technology journalist and editor with more than two decades of experience, Matt Vincent is Editor in Chief of Data Center Frontier.

Sign up for our eNewsletters
Get the latest news and updates
Image courtesy of Colocation America
Source: Image courtesy of Colocation America
Sponsored
Samantha Walters of Colocation America shares her thoughts on four trends she's seeing in the colocation space.
Blue Andy/Shutterstock.com
Source: Blue Andy/Shutterstock.com
Sponsored
Peter Huang, Global President - Data Center & Thermal Management at bp Castrol, explains why AI isn't just consuming more power, it's demanding better power systems.