News in the Channel - issue #16


The first option involves directing liquid to the server itself, using a room-based heat exchanger to reject heat back into the air. This modular system allows seamless integration without substantial changes to existing infrastructure. The second option introduces a Cooling Distribution Unit (CDU), directly circulating liquid from the server or GPU, connecting to a chilled water system. A third option is an interchangeable liquid- to-gas system. This approach incorporates a remote condenser on the roof or building, utilising gas-to-liquid heat exchangers for deployment flexibility. Ultimately, despite progress in liquid cooling, it is most likely that air-cooled and liquid-cooled solutions will co-exist. Even within liquid-cooled servers, elements necessitating air cooling persist, highlighting the nuanced nature of the evolving thermal management landscape.

may reflect a strategic pause to understand what technology changes and specifications are required. Challenges arise in finding the optimal operating conditions for CPUs and GPUs, with manufacturers defining specifications and clients striving to plan for a diverse technology landscape over the next five to 10 years. In this pursuit of future-ready design principles, clients encounter design pitfalls and challenges. The balance between CPU and GPU environments, coupled with defining optimal operating conditions, requires a meticulous approach to allow adaptability over an extended operational lifespan. As the industry grapples with these complexities, a holistic design ethos remains the compass guiding operators through the dynamic terrain of data centre evolution. Holistic design In this era dominated by AI, mobile and cloud technologies, and the advent of hybrid computing as the new norm, the importance of holistic design in data centres has never been more apparent. The evolution in the fungibility of workloads in the realm of AI signals a paradigm shift, recognising that workloads are no longer static, but dynamic, ever-changing entities. Navigating this dynamic landscape requires a holistic approach. Data centre architects, faced with the challenges of climate change, surging power requirements and heightened heat generation, are at the forefront of this transformative journey. In embracing a holistic design philosophy, data centres can position themselves to not only meet, but thrive in the face of the burgeoning demands of the AI-driven era. Sustainability and efficiency become the bedrock of operations, ensuring that data centres lead the charge in an era defined by growth and technological innovation.

Cooling distribution unit

Ensuring a holistic approach In the face of the ever-evolving and

rapidly changing landscape of data centre infrastructure, a holistic design approach emerges as the cornerstone for operators aiming to future-proof their operations and enable compatibility with new technologies and emerging waves of demand. The key to success lies in involving all stakeholders, recognising the importance of collaboration and communication across diverse disciplines. Engaging not only power and cooling specialists but also those responsible for facility management, storage and technology deployment, fosters a comprehensive understanding of the data centre’s intricate requirements.

Even within

liquid-cooled servers, elements necessitating air cooling persist, highlighting the nuanced nature of the evolving thermal management landscape.

As data centres embrace denser configurations and rapidly evolving

technology, the holistic approach extends to decision-making timelines. While operators may be inclined to defer decisions to the final stages of design, a balance must be struck to avoid risks associated with delayed investments and potential loss of market share. Holistic design, therefore, involves streamlining decision-making processes while considering lead times and involving stakeholders at every stage. In a dialogue with industry experts, the importance of technology interchangeability surfaces as a critical consideration for clients. In some areas we have seen a slowdown in direct deployments by hyperscalers, which


Powered by