How Can Data Center Managers Handle Explosive Growth Driven By AI and ML?

April 5, 2024
Marc Caiola, Vice President of Global Data Solutions at nVent, explains how leveraging the right cooling and power technologies can help data centers manage the growth of AI and ML.

Growing focus on generative Artificial Intelligence (AI) and Machine Learning (ML) has brought AI into the forefront of public conversation, but the capabilities of AI are vast. Today, we see large enterprises launching their own AI solutions and figuring out how to successfully integrate AI into business is top of mind for management. Supply chain management, research and analysis, and product marketing are just a few examples of how AI will be adopted to drive value for businesses and customers. AI will also see increased adoption in the healthcare, eMobility and energy generation and power industries as technology improves and key stakeholders grow more comfortable with its adoption.

All these factors are driving an increase in demand for the AI industry, which is expected to grow significantly in the coming decade.

New Technology Means New Demands

As data center managers know, it takes an extreme amount of data processing to deliver the results that users of AI and ML applications have come to expect and these applications are driven by high-performance chips on the cutting edge of IT development.

These advanced chips use a lot of power to run and produce more heat than less sophisticated applications. Data center managers have to deal with these high heat loads while still being able to scale their operations to meet demand. Scaling capabilities cannot always depend on more physical space—data center managers and engineers often have to solve the technical problem of fitting more and hotter servers into the same spaces. They also have to maintain 24/7 uptime: the needs of AI applications will not pause for a data center renovation.

Additionally, the industry is facing increased scrutiny over power use, so data center managers need to be especially conscious about how they are using electricity. Sustainability has always been a conversation in the data center industry, but this increased attention will create even more conversation around PUE and power management.

A Shift in Cooling Approach

Next-generation chips and other AI infrastructure need more than traditional cooling methods to keep them from overheating. Data centers may implement solutions that remove extra heat by increasing air volume or reducing air inlet temperatures, which can be inefficient and costly to run. When air cooling systems have to work harder to maintain optimal temperatures, facilities can also face equipment breakdowns, unexpected shutdowns and increased energy costs. For many data centers, using liquid cooling technologies can offer better performance while reducing energy use and helping data centers operate more sustainably. For the most advanced applications, liquid cooling is the only possible option. By using liquid cooling technologies in the right way, data center managers can greatly improve Power Usage Effectiveness (PUE), even in applications where they are using next-generation IT.

Liquid cooling can help data centers increase capacity while maintaining efficient space and energy use. It also can offer a favorable return on investment and lower the total cost of ownership for data center facilities. Liquid cooling systems provide an effective solution for achieving required temperature parameters and reducing the energy consumption of cooling systems. Liquid provides a much greater heat transfer capacity than air. This helps liquid cooling increase power usage effectiveness, managing heat loads effectively, reducing energy costs and contributing to environmental sustainability.   

Solutions at Scale

Liquid cooling does not have to be a comprehensive solution. Data centers can choose to cool a single rack or a few racks that run AI and machine learning applications without having to build entire data halls that are liquid cooled and support many racks of equipment that use high performance computing solutions.

However, when applying these partial solutions, it is vital to understand future business plans. Using specific cooling solutions for a particular problem is useful, but because of cost, energy efficiency and other factors, a solution for one problem may not work for another. As with all data center projects, different challenges need different solutions, and a universal approach rarely works.

With the growth in demand for high performance computing driven by the expansion of AI, data center managers need to have a plan in place to scale their cooling solutions. This may mean planning next generation data centers to be fully liquid cooled or exploring hybrid liquid to air solutions such as rear door cooling or direct-to-chip CDUs that bring liquid cooling to the rack and chip level while operating within air-cooled infrastructure.

The biggest advantage that planning for the future and understanding IT workloads will bring is the realization that almost all potential cooling solutions can be built out in combinations, allowing data center managers to match their power and cooling capabilities with shifting demands. The key to sustainable growth is a variety of flexible options for supporting the next-generation equipment. Liquid cooling technologies help drive that flexibility.

Power Management for AI

Power distribution is another critical technology for managing AI and ML workloads. Smart power distribution units (PDUs) are equipped with technology to distribute and monitor power usage to multiple devices within a data center and provide alerts in the event of power surges or other issues. 

The remote monitoring and control capabilities of smart PDUs can increase energy efficiency and reduce the risk of downtime. Input metering, for instance, allows power coming into a PDU to be remotely monitored, reducing the risk of overloading PDUs and causing breakers to trip. This monitoring can also help ensure that PDUs are not getting too close to breaker levels, allowing data center operators to remotely mitigate potential issues before they occur. 

Some PDUs also are equipped with outlet metering, where monitoring and control technology is applied not only to the PDU level but at the level of each individual outlet/power connection on the PDU. This technology can help operators better understand specific device power usage and compare efficiencies between different technologies. It can also identify underutilized or “zombie” equipment that is not in use, but still drawing significant power. Being able to remotely identify this equipment and turn it off allows data center managers to make sure they only are using the power that they need.

The opportunity for the data center industry brought about by the growth of AI and ML comes with challenges. By leveraging the right cooling and power technologies, data center managers can improve performance, drive sustainability and scale operations appropriately to meet the growing needs of their customers.

About the Author

Marc Caiola

Marc Caiola is the Vice President of Global Data Solutions at nVent.  Marc has over 30 years of technology and business leadership experience in Defense and ICT-Data and Communications Industries. Prior to joining nVent in 2018, Marc served in various marketing leadership roles within the Enclosures Strategic Business Unit at Pentair from 2007, and various leadership roles at Greco Systems, All Optical Networks and Comp Optics from 1996-2007. Marc served for 10 years in the United States Navy before entering the private sector in 1996. Marc holds electronics and computer science diplomas from the Chester County Technical College and United States Naval Training Center, a B.S. Business/Marketing from the University of Phoenix, and he completed executive development coursework from the University of Minnesota, Carlson School of Management.

Sponsored Recommendations

Get Utility Project Solutions

Lightweight, durable fiberglass conduit provides engineering benefits, performance and drives savings for successful utility project outcomes.

Guide to Environmental Sustainability Metrics for Data Centers

Unlock the power of Environmental, Social, and Governance (ESG) reporting in the data center industry with our comprehensive guide, proposing 28 key metrics across five categories...

The AI Disruption: Challenges and Guidance for Data Center Design

From large training clusters to small edge inference servers, AI is becoming a larger percentage of data center workloads. Learn more.

A better approach to boost data center capacity – Supply capacity agreements

Explore a transformative approach to data center capacity planning with insights on supply capacity agreements, addressing the impact of COVID-19, the AI race, and the evolving...


The Competitive Edge of Enterprise Edge

Brett Lindsey, CEO ark data centers, explains why edge data centers will become increasingly vital, making them the cornerstone of digital transformation strategies for enterprises...

White Papers

Get the full report.

Phoenix Data Center Market

June 28, 2022
Phoenix has historically drawn data center operator’s looking for an alternative to California’s higher costs and natural disaster risks, but today the market is a destination...