AMD and HPE are significantly expanding their collaboration, unveiling the AMD “Helios” rack-scale AI architecture. This initiative aims to deliver an open, full-stack AI platform engineered for large-scale AI workloads, marking a critical step in the evolution of enterprise AI infrastructure. According to the announcement, HPE will be among the first OEMs to adopt this new architecture, integrating its own networking solutions to enhance connectivity.
The "Helios" platform represents a cohesive AMD HPE AI infrastructure solution, combining AMD EPYC CPUs, AMD Instinct GPUs, AMD Pensando advanced networking, and the AMD ROCm open software stack. This integration is designed to simplify the deployment of massive AI clusters, promising faster time to solution and greater flexibility across diverse environments like research, cloud, and enterprise. The stated performance of up to 2.9 exaFLOPS of FP4 per rack with AMD Instinct MI455X GPUs and next-generation AMD EPYC “Venice” CPUs highlights its ambition for top-tier AI and HPC capabilities.
HPE’s role extends beyond mere adoption; it involves a strategic integration of purpose-built HPE Juniper Networking switches, developed in collaboration with Broadcom. These switches are crucial for delivering the high-bandwidth, low-latency connectivity essential for massive AI clusters. Leveraging the Ultra Accelerator Link over Ethernet (UALoE) standard, this networking component reinforces the commitment to open, standards-based technologies, a significant differentiator in a market often characterized by proprietary solutions. This focus on open standards could alleviate concerns about vendor lock-in, offering greater flexibility for customers.
Advancing HPC and Sovereign AI
Beyond the rack-scale architecture, the collaboration is also powering "Herder," a new supercomputer for the High-Performance Computing Center Stuttgart (HLRS) in Germany. Built on the HPE Cray GX5000 platform, Herder will feature AMD Instinct MI430X GPUs and next-generation AMD EPYC “Venice” CPUs. This deployment underscores the strategic importance of the AMD HPE AI infrastructure partnership in advancing high-performance computing and sovereign AI research across Europe. The system is designed to support both traditional HPC applications and emerging machine learning workflows, providing a versatile tool for scientific discovery and industrial innovation.
The emphasis on an "open" architecture, from the OCP Open Rack Wide design to the ROCm software stack and UALoE standard, is a deliberate strategic move. This approach directly challenges more closed ecosystems prevalent in the AI hardware space, offering enterprises and cloud providers a more adaptable and potentially cost-effective path to scaling their AI operations. By championing open standards, AMD and HPE are positioning themselves as enablers of broader innovation, reducing barriers to entry for developers and researchers. This commitment could foster a more vibrant and competitive AI infrastructure landscape.
This expanded AMD HPE AI infrastructure collaboration signals a mature phase in the AI arms race, moving beyond individual component performance to integrated, rack-scale solutions. The focus on openness, scalability, and high-performance networking addresses critical pain points for organizations deploying large-scale AI. As the industry grapples with increasing AI demands and the complexities of infrastructure management, this partnership offers a compelling vision for the future of accessible and powerful AI computing.



